davila7
GitHub profile for davila741 skills
davila7 / matchms
Facilitates mass spectrometry data processing and analysis, enabling users to import, filter, and compare spectral data efficiently.
davila7 / crewai-multi-agent
Facilitates the orchestration of autonomous AI agents for collaborative problem-solving in complex tasks and workflows.
davila7 / langchain
LangChain enables the development of LLM-powered applications with agents and RAG, supporting multiple providers and integrations.
davila7 / llamaindex
LlamaIndex is a data framework for building LLM applications, enabling document ingestion, indexing, and querying for enhanced knowledge retrieval.
davila7 / huggingface-accelerate
Simplifies distributed training in PyTorch with a unified API for various frameworks, enabling easy multi-GPU and mixed precision setups.
davila7 / deepspeed
Provides expert guidance for distributed training with DeepSpeed, optimizing large-scale deep learning models efficiently.
davila7 / training-llms-megatron
Facilitates training of large language models using NVIDIA Megatron-Core with advanced parallelism for optimal GPU efficiency.
davila7 / pytorch-lightning
Facilitates scalable and efficient training of PyTorch models with minimal boilerplate and built-in best practices.
davila7 / long-context
Extends transformer model context windows using advanced techniques like RoPE and ALiBi for processing long documents efficiently.
davila7 / model-merging
Enables merging of multiple fine-tuned models without retraining, enhancing performance and reducing costs for specialized AI applications.
davila7 / moe-training
Enables efficient training of Mixture of Experts models using DeepSpeed or HuggingFace, optimizing compute costs and performance.
davila7 / evaluating-code-models
Evaluates code generation models using benchmarks like HumanEval and MBPP, providing insights into coding abilities and quality.
davila7 / nemo-evaluator-sdk
Evaluates LLMs using 100+ benchmarks with scalable execution on Docker and Slurm HPC for reproducible results.
davila7 / gepetto
Facilitates comprehensive implementation planning through research and stakeholder engagement, ensuring thorough pre-implementation analysis.
davila7 / llama-cpp
Enables efficient LLM inference on non-NVIDIA hardware, optimizing performance for edge deployment and Apple Silicon.
davila7 / sglang
Enables fast structured generation and serving for LLMs, optimizing inference with RadixAttention for efficient workflows.
davila7 / tensorrt-llm
Optimizes LLM inference using NVIDIA TensorRT for high throughput and low latency on NVIDIA GPUs, enhancing production deployment efficiency.
davila7 / serving-llms-vllm
Optimizes LLM serving with high throughput using vLLM, enabling efficient deployment and inference for production APIs.
davila7 / jira
Facilitates interaction with Jira for managing issues, tickets, and sprints through natural language commands and backend detection.
davila7 / loki-mode
Automates startup processes by orchestrating specialized agents across various domains, enabling zero human intervention from PRD to deployment.