Homeā€ŗCompaniesā€ŗdeepsilicon

deepsilicon

Software and hardware to run neural networks faster and cheaper

deepsilicon
Founded:2024
Team Size:2
Location:San Francisco
Group Partner:Diana Hu

Active Founders

Abhinav Reddy, Founder

Computer Science and Electrical Engineering. Focused on creating simple software and fast hardware for running neural networks.
Abhinav Reddy
Abhinav Reddy
deepsilicon

Alexander Nanda, Founder

Dartmouth College Computer Science and Physics, building software and hardware for fast neural nets
Alexander Nanda
Alexander Nanda
deepsilicon

Company Launches

šŸ‘„ The Team
Hi everyone, weā€™re Abhi and Alex, and weā€™re running deepsilicon!

šŸ“‰ The Problem
Transformer-based models have become increasingly crucial in various industries, from natural language processing to Vision Language Action models for robotics. However, the deployment and operation of these models, particularly those exceeding a few billion parameters, present significant challenges regarding hardware capabilities, energy consumption, and operational costs.
Traditional approaches to this problem typically fall into two categories:

  1. Utilize massive, power-hungry GPU clusters to distribute the computational load.
  2. Compromise on model size and capabilities to fit within existing hardware constraints on the edge.

Both of these approaches have significant drawbacks. GPU clusters are expensive to acquire and operate, with substantial energy costs and complex cooling requirements. They also introduce latency issues due to inter-device communication and canā€™t be deployed on the edge. On the other hand, compromising on model size can limit the AI's capabilities and potential applications, putting organizations at a competitive disadvantage.


šŸ“ˆ The Solution
We help eliminate the need for inefficient distributed computing and compromised model capabilities by providing a full-stack system where we run transformer-based models on a single chip, including existing hardware. Our solution can run on a custom ASIC, dramatically reducing power consumption and operational costs.

https://www.youtube.com/watch?v=MctVUhuXgeA


Here's why this is a game-changer:

  • Immediate Deployment:Ā Your large-scale AI model can be operational on a single chip, eliminating the need for complex distributed setups. This means you can leverage the full power of multi-billion parameter models right from the start.
  • Customization:Ā The chiplet architecture allows for near-infinite customization, enabling hardware/software co-design tailored to specific client needs.
  • Efficiency:Ā Even on existing Nvidia hardware, our software provides a 5x reduction in memory usage
  • Ease of Use:Ā Developers can switch their linear layer with our optimized version, dramatically simplifying the integration process and democratizing access to large-scale AI capabilities.

If youā€™re a YC company, we offer a 50% discount to help you train your model and deploy it on the device or the cloud!

šŸ™ How You Can Help
We are looking for connections and collaborations to drive our mission forward. If you or anyone in your network is interested in learning more or helping out, we want to hear from you! Weā€™re specifically looking to connect with:

  • Robotics, Drone, and Autonomous Vehicle Companies
  • People looking to train their own foundation model
  • Experts in edge AI deployment

If thatā€™s you or someone you know, please email us atĀ founders@deepsilicon.net!