Showing 131 open source projects for "python programming language"

View related business solutions
  • Try Google Cloud Risk-Free With $300 in Credit Icon
    Try Google Cloud Risk-Free With $300 in Credit

    No hidden charges. No surprise bills. Cancel anytime.

    Use your credit across every product. Compute, storage, AI, analytics. When it runs out, 20+ products stay free. You only pay when you choose to.
    Start Free
  • Earn up to 16% annual interest with Nexo. Icon
    Earn up to 16% annual interest with Nexo.

    Access competitive interest rates on your digital assets.

    Generate interest, borrow against your crypto, and trade a range of cryptocurrencies — all in one platform. Geographic restrictions, eligibility, and terms apply.
    Get started with Nexo.
  • 1
    MetaTransformer

    MetaTransformer

    Meta-Transformer for Unified Multimodal Learning

    We're thrilled to present OneLLM, an ensembling Meta-Transformer framework with Multimodal Large Language Models, which performs multimodal joint training, supports more modalities including fMRI, Depth, and Normal Maps, and demonstrates very impressive performances on 25 benchmarks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 2
    Transformer Reinforcement Learning X

    Transformer Reinforcement Learning X

    A repo for distributed training of language models with Reinforcement

    trlX is a distributed training framework designed from the ground up to focus on fine-tuning large language models with reinforcement learning using either a provided reward function or a reward-labeled dataset. Training support for Hugging Face models is provided by Accelerate-backed trainers, allowing users to fine-tune causal and T5-based language models of up to 20B parameters, such as facebook/opt-6.7b, EleutherAI/gpt-neox-20b, and google/flan-t5-xxl. For models beyond 20B parameters,...
    Downloads: 5 This Week
    Last Update:
    See Project
  • 3

    Lumi-HSP

    This is an AI language model that can predict Heart failure or stroke

    Using thsi AI model, you can predict the chances of heart stroke and heart failure. HIGLIGHTS : 1. Accuracy of this model is 95% 2. This model uses the powerful Machine Learning algorithm "GradientBoosting" for predicting the outcomes. 3. An easy to use model and accessible to everyone.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 4
    picoGPT

    picoGPT

    An unnecessarily tiny implementation of GPT-2 in NumPy

    picoGPT is a minimal implementation of the GPT-2 language model designed to demonstrate how transformer-based language models work at a conceptual level. The repository focuses on educational clarity rather than production performance, implementing the core components of the GPT architecture in a concise and readable way. It allows users to understand how tokenization, transformer layers, attention mechanisms, and autoregressive text generation operate in modern large language models. The...
    Downloads: 0 This Week
    Last Update:
    See Project
  • Gemini 3 and 200+ AI Models on One Platform Icon
    Gemini 3 and 200+ AI Models on One Platform

    Access Google's best plus Claude, Llama, and Gemma. Fine-tune and deploy from one console.

    Build generative AI apps with Vertex AI. Switch between models without switching platforms.
    Start Free
  • 5
    MLPACK is a C++ machine learning library with emphasis on scalability, speed, and ease-of-use. Its aim is to make machine learning possible for novice users by means of a simple, consistent API, while simultaneously exploiting C++ language features to provide maximum performance and flexibility for expert users. * More info + downloads: https://mlpack.org * Git repo: https://github.com/mlpack/mlpack
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    CodeContests

    CodeContests

    Large dataset of coding contests designed for AI and ML model training

    CodeContests, developed by Google DeepMind, is a large-scale competitive programming dataset designed for training and evaluating machine learning models on code generation and problem solving. This dataset played a central role in the development of AlphaCode, DeepMind’s model for solving programming problems at a human-competitive level, as published in Science. CodeContests aggregates problems and human-written solutions from multiple programming competition platforms, including AtCoder,...
    Downloads: 2 This Week
    Last Update:
    See Project
  • 7
    Mars Framework

    Mars Framework

    Mars is a tensor-based unified framework for large-scale data

    Mars is a distributed computing framework designed to scale scientific computing and data science workloads across large clusters while preserving the familiar programming interfaces of common Python libraries. The project provides a tensor-based execution model that extends the capabilities of tools such as NumPy, pandas, and scikit-learn so that large datasets can be processed in parallel without rewriting code for distributed environments. Its architecture automatically divides large computational tasks into smaller chunks that can be executed across multiple nodes in a cluster, allowing complex analytics, machine learning workflows, and data transformations to run efficiently at scale. ...
    Downloads: 6 This Week
    Last Update:
    See Project
  • 8
    d2l-zh

    d2l-zh

    Chinese-language edition of Dive into Deep Learning

    d2l‑zh is the Chinese-language edition of Dive into Deep Learning, an interactive, open‑source deep learning textbook that combines code, math, and explanatory text. It features runnable Jupyter notebooks compatible with multiple frameworks (e.g., PyTorch, MXNet, TensorFlow), comprehensive theoretical analysis, and exercises. Widely adopted in over 70 countries and used by more than 500 universities for teaching deep learning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    DialoGPT

    DialoGPT

    Large-scale pretraining for dialogue

    DialoGPT is an open-source conversational language model developed by Microsoft Research for generating natural dialogue responses using large-scale transformer architectures. The system is built on the GPT-2 architecture and is designed specifically for multi-turn conversation tasks, enabling machines to produce coherent responses during interactive dialogue. The model was trained on a massive dataset of approximately 147 million conversational exchanges extracted from Reddit discussion...
    Downloads: 0 This Week
    Last Update:
    See Project
  • MongoDB Atlas runs apps anywhere Icon
    MongoDB Atlas runs apps anywhere

    Deploy in 115+ regions with the modern database for every enterprise.

    MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
    Start Free
  • 10
    Pattern

    Pattern

    Web mining module for Python, with tools for scraping

    Pattern is an open-source Python library that provides tools for web mining, natural language processing, machine learning, and network analysis. The project integrates multiple capabilities into a single framework that allows developers to collect, process, and analyze textual data from the web. It includes modules for web scraping and crawling that can retrieve information from sources such as social media platforms, search engines, and online knowledge bases.
    Downloads: 11 This Week
    Last Update:
    See Project
  • 11
    PromptSource

    PromptSource

    Toolkit for creating, sharing and using natural language prompts

    PromptSource is a toolkit for creating, sharing and using natural language prompts. Recent work has shown that large language models exhibit the ability to perform reasonable zero-shot generalization to new tasks. For instance, GPT-3 demonstrated that large language models have strong zero- and few-shot abilities. FLAN and T0 then demonstrated that pre-trained language models fine-tuned in a massively multitask fashion yield even stronger zero-shot performance. A common denominator in these...
    Downloads: 8 This Week
    Last Update:
    See Project
  • 12
    OpenPrompt

    OpenPrompt

    An Open-Source Framework for Prompt-Learning

    Prompt-learning is the latest paradigm to adapt pre-trained language models (PLMs) to downstream NLP tasks, which modifies the input text with a textual template and directly uses PLMs to conduct pre-trained tasks. OpenPrompt is a library built upon PyTorch and provides a standard, flexible and extensible framework to deploy the prompt-learning pipeline. OpenPrompt supports loading PLMs directly from huggingface transformers. In the future, we will also support PLMs implemented by other...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 13
    Apache MXNet (incubating)

    Apache MXNet (incubating)

    A flexible and efficient library for deep learning

    Apache MXNet is an open source deep learning framework designed for efficient and flexible research prototyping and production. It contains a dynamic dependency scheduler that automatically parallelizes both symbolic and imperative operations. On top of this is a graph optimization layer, overall making MXNet highly efficient yet still portable, lightweight and scalable.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 14
    CodeSearchNet

    CodeSearchNet

    Datasets, tools, and benchmarks for representation learning of code

    ...The dataset currently covers several widely used programming languages, including Python, JavaScript, Ruby, Go, Java, and PHP. In addition to the dataset itself, the repository includes baseline models, evaluation tools, and instructions for building code retrieval systems that can map user queries to relevant code snippets.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 15
    Machine Learning PyTorch Scikit-Learn

    Machine Learning PyTorch Scikit-Learn

    Code Repository for Machine Learning with PyTorch and Scikit-Learn

    ...The first half of the book introduces readers to machine learning using scikit-learn, the defacto approach for working with tabular datasets. Then, the second half of this book focuses on deep learning, including applications to natural language processing and computer vision.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 16
    Hugging Face Transformer

    Hugging Face Transformer

    CPU/GPU inference server for Hugging Face transformer models

    Optimize and deploy in production Hugging Face Transformer models in a single command line. At Lefebvre Dalloz we run in-production semantic search engines in the legal domain, in the non-marketing language it's a re-ranker, and we based ours on Transformer. In that setup, latency is key to providing a good user experience, and relevancy inference is done online for hundreds of snippets per user query. Most tutorials on Transformer deployment in production are built over Pytorch and FastAPI....
    Downloads: 1 This Week
    Last Update:
    See Project
  • 17
    Awesome Decision Tree Papers

    Awesome Decision Tree Papers

    A collection of research papers on decision, classification, etc.

    A collection of research papers on decision, classification and regression trees with implementations.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    FARM

    FARM

    Fast & easy transfer learning for NLP

    FARM makes Transfer Learning with BERT & Co simple, fast and enterprise-ready. It's built upon transformers and provides additional features to simplify the life of developers: Parallelized preprocessing, highly modular design, multi-task learning, experiment tracking, easy debugging and close integration with AWS SageMaker. With FARM you can build fast proofs-of-concept for tasks like text classification, NER or question answering and transfer them easily into production. Easy fine-tuning...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    Kashgari

    Kashgari

    Kashgari is a production-level NLP Transfer learning framework

    Kashgari is a simple and powerful NLP Transfer learning framework, build a state-of-art model in 5 minutes for named entity recognition (NER), part-of-speech tagging (PoS), and text classification tasks.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    AliceMind

    AliceMind

    ALIbaba's Collection of Encoder-decoders from MinD

    This repository provides pre-trained encoder-decoder models and its related optimization techniques developed by Alibaba's MinD (Machine IntelligeNce of Damo) Lab. Pre-trained models for natural language understanding (NLU). We extend BERT to a new model, StructBERT, by incorporating language structures into pre-training. Specifically, we pre-train StructBERT with two auxiliary tasks to make the most of the sequential order of words and sentences, which leverage language structures at the...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 21
    SRU

    SRU

    Training RNNs as Fast as CNNs

    Common recurrent neural architectures scale poorly due to the intrinsic difficulty in parallelizing their state computations. In this work, we propose the Simple Recurrent Unit (SRU), a light recurrent unit that balances model capacity and scalability. SRU is designed to provide expressive recurrence, enable highly parallelized implementation, and comes with careful initialization to facilitate the training of deep models. We demonstrate the effectiveness of SRU on multiple NLP tasks. SRU...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 22
    Couler

    Couler

    Unified Interface for Constructing and Managing Workflows

    Couler is a system designed for unified machine learning workflow optimization in the cloud. Couler endeavors to provide a unified interface for constructing and optimizing workflows across various workflow engines, such as Argo Workflows, Tekton Pipelines, and Apache Airflow. Couler enhances workflow efficiency through features like Autonomous Workflow Construction, Automatic Artifact Caching Mechanisms, Big Workflow Auto Parallelism Optimization, and Automatic Hyperparameters Tuning.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 23
    PORORO

    PORORO

    Platform of neural models for natural language processing

    pororo performs Natural Language Processing and Speech-related tasks. It is easy to solve various subtasks in the natural language and speech processing field by simply passing the task name. Recognized speech sentences using the trained model. Currently English, Korean and Chinese support. Get vector or find similar words and entities from pretrained model using Wikipedia.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 24
    PaddlePaddle models

    PaddlePaddle models

    Pre-trained and Reproduced Deep Learning Models

    Pre-trained and Reproduced Deep Learning Models ("Flying Paddle" official model library, including a variety of academic frontier and industrial scene verification of deep learning models) Flying Paddle's industrial-level model library includes a large number of mainstream models that have been polished by industrial practice for a long time and models that have won championships in international competitions; it provides many scenarios for semantic understanding, image classification,...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 25
    NLP Architect

    NLP Architect

    A model library for exploring state-of-the-art deep learning

    NLP Architect is an open-source Python library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing and Natural Language Understanding neural networks. The library includes our past and ongoing NLP research and development efforts as part of Intel AI Lab. NLP Architect is designed to be flexible for adding new models, neural network components, data handling methods, and for easy training and running models. ...
    Downloads: 0 This Week
    Last Update:
    See Project
MongoDB Logo MongoDB