Performance
Model size: 1.5B - 70B parameters
Inference speed: > 100 tokens/s
Inference speed: > 100 tokens/s
Innovative core architecture
Compute-in-memory technology
Dataflow computing
Outstanding cost-effectiveness and power efficiency
Edge LLM SoCs