Mostrando 5,781 - 5,800 Resultados de 9,480 Para Buscar '"Sage"', tiempo de consulta: 0.09s Limitar resultados
  1. 5781
    Publicado 2002
    Revista digital
  2. 5782
    Publicado 2016
    Revista digital
  3. 5783
    Publicado 2002
    Revista digital
  4. 5784
    Publicado 2001
    Revista digital
  5. 5785
    Publicado 1986
    Revista digital
  6. 5786
    Publicado 1959
    Revista digital
  7. 5787
    Publicado 2001
    Revista digital
  8. 5788
    Publicado 1996
    Revista digital
  9. 5789
    Publicado 1955
    Revista digital
  10. 5790
  11. 5791
  12. 5792
    Publicado 2024
    Tabla de Contenidos: “…Quantization with GPTQ and EXL2 -- Other quantization techniques -- Summary -- References -- Chapter 9: RAG Inference Pipeline -- Understanding the LLM twin's RAG inference pipeline -- Exploring the LLM twin's advanced RAG techniques -- Advanced RAG pre-retrieval optimizations: query expansion and self-querying -- Query expansion -- Self-querying -- Advanced RAG retrieval optimization: filtered vector search -- Advanced RAG post-retrieval optimization: reranking -- Implementing the LLM twin's RAG inference pipeline -- Implementing the retrieval module -- Bringing everything together into the RAG inference pipeline -- Summary -- References -- Chapter 10: Inference Pipeline Deployment -- Criteria for choosing deployment types -- Throughput and latency -- Data -- Understanding inference deployment types -- Online real-time inference -- Asynchronous inference -- Offline batch transform -- Monolithic versus microservices architecture in model serving -- Monolithic architecture -- Microservices architecture -- Choosing between monolithic and microservices architectures -- Exploring the LLM Twin's inference pipeline deployment strategy -- The training versus the inference pipeline -- Deploying the LLM Twin service -- Implementing the LLM microservice using AWS SageMaker -- What are Hugging Face's DLCs? -- Configuring SageMaker roles -- Deploying the LLM Twin model to AWS SageMaker -- Calling the AWS SageMaker Inference endpoint -- Building the business microservice using FastAPI -- Autoscaling capabilities to handle spikes in usage -- Registering a scalable target -- Creating a scalable policy -- Minimum and maximum scaling limits -- Cooldown period -- Summary -- References -- Chapter 11: MLOps and LLMOps -- The path to LLMOps: Understanding its roots in DevOps and MLOps -- DevOps -- The DevOps lifecycle -- The core DevOps concepts -- MLOps…”
    Libro electrónico
  13. 5793
    Publicado 2007
    Acceso restringido con credenciales, usuarios UPSA
    Libro electrónico
  14. 5794
    Publicado 2010
    Acceso restringido con credenciales, usuarios UPSA
    Libro electrónico
  15. 5795
    Publicado 1984
    Libro
  16. 5796
    por Hoover, Stewart M.
    Publicado 1989
    Capítulo
  17. 5797
    Publicado 1980
    Revista
  18. 5798
    Publicado 1996
    Libro
  19. 5799
    Publicado 1997
    Libro
  20. 5800
    por Klein, David M., 1943-
    Publicado 1996
    Libro