Found 10 AI tools
Click any tool to view details
Intel® Core™ Extreme 200 Series Desktop Processors are the first AI PC processors for desktop platforms, bringing enthusiasts an excellent gaming experience and industry-leading computing performance while significantly reducing power consumption. These processors feature up to eight next-generation performance cores (P-cores) and up to 16 next-generation energy efficiency cores (E-cores), delivering up to 14% performance improvements in multi-threaded workloads compared to the previous generation. These processors are the first desktop processors to feature a Neural Processing Unit (NPU) for enthusiasts with built-in Xe GPU to support the most advanced media capabilities.
DataGemma is the world's first open model designed to help solve the problem of AI hallucinations through massive amounts of real-world statistics from Google's data sharing platform. These models enhance the factuality and reasoning capabilities of the language model through two different methods, thereby reducing hallucinations and improving the accuracy and reliability of AI. The launch of the DataGemma model is an important advancement in AI technology in improving data accuracy and reducing the spread of misinformation. It is of great significance to researchers, policymakers, and ordinary users.
HumanPlus is a research project aimed at training humanoid robots to achieve autonomous skill learning by imitating human movements. This project trains low-level policies through simulated reinforcement learning and applies these policies to the real world to achieve real-time tracking of human body and hand movements. Through shadow simulation technology, operators can remotely operate the robot to collect whole-body data for learning different tasks. In addition, through behavioral cloning technology, robots can imitate human skills and complete various tasks.
AutoMathText is an extensive and carefully curated dataset containing approximately 200GB of mathematical text. Each piece of content in the dataset is independently selected and scored by Qwen, the most advanced open source language model, ensuring high standards of relevance and quality. This dataset is particularly suitable for promoting advanced research at the intersection of mathematics and artificial intelligence, as an educational tool for learning and teaching complex mathematical concepts, and as a basis for developing and training AI models that specifically process and understand mathematical content.
Mamba Bytes is a token-free language model that learns directly from raw bytes, eliminating the bias of sub-word tokenization. It operates on bytes, but results in significantly longer sequences, and the standard autoregressive Transformer scales poorly in this case. We autoregressively trained Mamba bytes on byte sequences, which is a label-free adaptation of the Mamba state space model. Our experiments show that Mambabytes is computationally efficient compared to other byte-level models. We also found that Mambabyte performs well against and even exceeds the performance of the state-of-the-art Subword Transformer. Furthermore, MambaByte has faster speed than Transformer during inference due to the linear expansion of length. Our findings confirm the feasibility of mambabytes in enabling markup-free language modeling.
BiTA is a bidirectional tuning method for large language models that accelerates large language models through simplified semi-autoregressive generation and draft verification. As a lightweight plug-in module, BiTA can seamlessly improve the inference efficiency of existing large language models without requiring additional auxiliary models or incurring significant additional memory costs. After applying BiTA, LLaMA-2-70B-Chat achieved a 2.7x acceleration on the MT-Bench benchmark. Extensive experiments confirm that our approach surpasses state-of-the-art acceleration techniques.
E^2-LLM is an efficient and extremely extended large language model method. It achieves effective support for long context tasks by requiring only one training process and greatly reducing computational costs. This method adopts RoPE position embedding and introduces two different enhancement methods aimed at making the model more robust during inference. Comprehensive experimental results on multiple benchmark datasets demonstrate the effectiveness of E^2-LLM on challenging long context tasks.
This is an efficient LLM inference solution implemented on Intel GPUs. By simplifying the LLM decoder layer, using a segmented KV caching strategy and a custom Scaled-Dot-Product-Attention kernel, the solution achieves up to 7x lower token latency and 27x higher throughput on Intel GPUs compared to the standard HuggingFace implementation. Please refer to the official website for detailed functions, advantages, pricing and positioning information.
Astraios is a platform that provides fine-tuning for large language models. It provides efficient fine-tuning methods with multiple parameters and model selections of various sizes. Users can fine-tune large-scale language models on this platform and obtain the best cost-performance balance. The platform also provides a wealth of models, data sets and documents to facilitate users to conduct related research and development. Pricing is flexible and suitable for user needs of different sizes.
Meta 4 WCS AI is a similarity engine based on Intel TF-iDF, which can provide users with key debugging information to help optimize business processes.
Explore other subcategories under other Other Categories
195 tools
178 tools
113 tools
102 tools
62 tools
61 tools
49 tools
45 tools
AI model inference training Hot other is a popular subcategory under 10 quality AI tools