Sublime
An inspiration engine for ideas
🍵What is AutoMQ
AutoMQ is a cloud-native, serverless reinvented Kafka that is easily scalable, manage-less and cost-effective.
đź”¶Why AutoMQ
AutoMQ is a cloud-native, serverless reinvented Kafka that is easily scalable, manage-less and cost-effective.
đź”¶Why AutoMQ
- Cloud Native : Built on cloud service. Every system design decision take cloud service's feature and billing items into consideration to offer best low-latency, scalable, reliable and cost-effective Kafka serv
AutoMQ • GitHub - AutoMQ/automq: A cloud native implementation for Apache Kafka, reducing your cloud infrastructure bill by up to 90%.
Quorum
quorum.chat
The Nemotron-3 8B family is available in the Azure AI Model Catalog, HuggingFace, and the NVIDIA AI Foundation Model hub on the NVIDIA NGC Catalog. It includes base, chat, and question-and-answer (Q&A) models that are designed to solve a variety of downstream tasks. Table 1 shows the full family of foundation models.
Model
Variant
Key Benefit
Ba... See more
Model
Variant
Key Benefit
Ba... See more
NVIDIA Technical Blog | News and tutorials for developers, data ...
Qwen-14B is the 14B-parameter version of the large language model series, Qwen (abbr. Tongyi Qianwen), proposed by Alibaba Cloud. Qwen-14B is a Transformer-based large language model, which is pretrained on a large volume of data, including web texts, books, codes, etc. Additionally, based on the pretrained Qwen-14B, we release Qwen-14B-Chat, a lar... See more
Qwen/Qwen-14B-Chat · Hugging Face
AWS Announces General Availability of Amazon Q, the Most Capable Generative AI-Powered Assistant for Accelerating Software Development and Leveraging Companies’ Internal Data
press.aboutamazon.com
2-5x faster 50% less memory local LLM finetuning
- Manual autograd engine - hand derived backprop steps.
- 2x to 5x faster than QLoRA. 50% less memory usage.
- All kernels written in OpenAI's Triton language.
- 0% loss in accuracy - no approximation methods - all exact.
- No change of hardware necessary. Supports NVIDIA GPUs since 2018+. Minimum CUDA Compute Cap
unslothai • GitHub - unslothai/unsloth: 5X faster 50% less memory LLM finetuning
🤖 Cutting-edge framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
- Why CrewAI
- Getting Started
- Key Features
- Examples
- Local Open Source Models
- CrewAI x AutoGen x ChatDev
- Contribution
- đź’¬ CrewAI Discord Community
- Hire Consulting
- Licen
joaomdmoura • GitHub - joaomdmoura/crewAI: Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
pair-preference-model-LLaMA3-8B by RLHFlow: Really strong reward model, trained to take in two inputs at once, which is the top open reward model on RewardBench (beating one of Cohere’s).
DeepSeek-V2 by deepseek-ai (21B active, 236B total param.): Another strong MoE base model from the DeepSeek team. Some people are questioning the very high MMLU sc... See more
DeepSeek-V2 by deepseek-ai (21B active, 236B total param.): Another strong MoE base model from the DeepSeek team. Some people are questioning the very high MMLU sc... See more