Guest



Sign inSignup
  • Home
  • Dashboard
  • Tools
  • Wiro AppsApps
  • Pricing

Welcome

HomeDashboardToolsWiro AppsApps
Use cases
Human Resources
Retail & E-commerce
Interior Design
Fashion AI
Creative Content Solutions
Sports & Fitness
GenAI Video Tools
PricingDocumentation
Guest



Sign inSignup

Task History

  • Runnings
  • Models
  • Trains

You don't have task yet.

Go to Tools

Welcome

deepseek-ai

  • Models(7)
View All

We couldn't find any matching results.

Too bad. Looks like we couldn't find any matching results for the search terms you've entered. Please try different search terms or criteria.

deepseek-math-7b-instruct

deepseek-ai/deepseek-math-7b-instruct is a 7 billion parameter AI model specifically designed to excel in mathematical reasoning and problem-solving tasks, providing accurate and detailed step-by-step solutions.

DeepSeek-R1-Distill-Qwen-1.5B

DeepSeek-R1-Distill-Qwen-1.5B is a distilled version of the DeepSeek-R1 model, designed to offer a balance between efficiency and performance. With 1.5 billion parameters, it leverages knowledge distillation techniques to retain the capabilities of larger models while optimizing for faster inference and reduced resource consumption. Ideal for applications requiring high-quality language understanding with lower computational overhead.

DeepSeek-R1-Distill-Qwen-7B

DeepSeek-R1-Distill-Qwen-7B is a distilled version of the DeepSeek-R1 model with 7 billion parameters, designed to provide high-quality language understanding while optimizing efficiency. Leveraging advanced knowledge distillation techniques, it retains the core capabilities of larger models with improved speed and lower resource consumption. This model is well-suited for tasks requiring robust natural language processing while maintaining cost-effective deployment.

DeepSeek-R1-Distill-Llama-8B

DeepSeek-R1-Distill-Llama-8B is a distilled version of the DeepSeek-R1 model based on the LLaMA architecture, featuring 8 billion parameters. It offers a balance between performance and efficiency by leveraging knowledge distillation techniques to reduce computational costs while maintaining high-quality language processing capabilities. This model is ideal for applications that require powerful text generation and understanding with optimized resource usage.

DeepSeek-R1-Distill-Qwen-14B

DeepSeek-R1-Distill-Qwen-14B is a distilled version of the DeepSeek-R1 model, featuring 14 billion parameters. Built on the Qwen architecture, it utilizes advanced knowledge distillation techniques to achieve a balance between high performance and computational efficiency. This model is well-suited for a wide range of natural language processing tasks, providing accurate and context-aware responses while optimizing resource consumption for deployment in production environments.

DeepSeek-R1-Distill-Qwen-32B

DeepSeek-R1-Distill-Qwen-32B is a distilled version of the DeepSeek-R1 model with 32 billion parameters, leveraging the Qwen architecture to deliver high-quality language understanding and generation. Through knowledge distillation, it retains the strengths of larger models while offering improved efficiency and reduced computational requirements. This model is ideal for large-scale AI applications that demand robust performance with optimized resource utilization.

Select Language

Logo of nvidia programLogo of nvidia program
Wiro AI brings machine learning easily accessible to all in the cloud.
  • WIRO
  • About
  • Careers
  • Contact
  • Language Language
  • Product
  • Tools
  • Pricing
  • Roadmap
  • Changelog
  • Status
  • Documentation
  • Introduction
  • Start Your First Project
  • Example Projects

2023 © Wiro.ai | Terms of Service & Privacy Policy