Prompt Cards

Micro-batching and latency trade-offs for streaming LLMs
Analyze this: Micro-batching and latency trade-offs for streaming LLMs
Tags: gpu, latency, streaming, micro-batching, LLM
Author: judy
Created at: 2025-11-05 11:15:00
Average Rating:
Total Ratings:
Energy-efficient LLM inference strategies
Analyze relevant market data that will benefit from: Energy-efficient LLM inference strategies
Tags: gpu, energy, efficiency, inference, LLM
Author: ivan
Created at: 2025-11-05 11:00:00
Average Rating:
Total Ratings:
Convert Hugging Face LLM to TensorRT and Triton
Convert Hugging Face LLM to TensorRT and Triton. Give clear instructions on how to do it.
Tags: gpu, TensorRT, Triton, deployment, ONNX
Author: heidi
Created at: 2025-11-05 10:45:00
Average Rating:
Total Ratings:
Multi-GPU sharding and parallelism strategy
Give guide for software engineers for: Multi-GPU sharding and parallelism strategy
Tags: gpu, distributed, sharding, parallelism, LLM
Author: grace
Created at: 2025-11-05 10:30:00
Average Rating:
Total Ratings:
Profiling LLM inference on GPU: commands and metrics
Give important information about: Europe's Winter Energy Outlook: Local and Regional Risks
Tags: gpu, profiling, inference, performance
Author: frank
Created at: 2025-11-05 10:15:00
Average Rating:
Total Ratings:
Mixed-precision training checklist for LLMs
List in detail about: Mixed-precision training checklist for LLMs
Tags: gpu, mixed-precision, training, stability
Author: eve
Created at: 2025-11-05 10:00:00
Average Rating:
Total Ratings:
Quantization pipeline for 70B models
Expalin in detail: Quantization pipeline for 70B models
Tags: gpu, quantization, LLM, model-compression
Author: dave
Created at: 2025-11-05 09:45:00
Average Rating:
Total Ratings:
Choosing GPUs for LLM inference at scale
Give detailed steps on:Urban Crime Trends in US Cities: Data-Driven Local Reporting
Tags: gpu, inference, hardware, capacity, cost
Author: carol
Created at: 2025-11-05 09:30:00
Average Rating:
Total Ratings:
CUDA kernel optimizations for attention
Create an engineer's guidebook for CUDA kernel optimizations for attention
Tags: gpu, kernels, attention, CUDA, performance
Author: bob
Created at: 2025-11-05 09:15:00
Average Rating:
Total Ratings:
GPU memory optimization for LLM fine-tuning
Give information that is useful to software enginners: GPU memory optimization for LLM fine-tuning
Tags: gpu, LLM, training, memory, optimization
Author: alice
Created at: 2025-11-05 09:00:00
Average Rating:
Total Ratings:
Border Tensions and Migration Pressures in Eastern Europe: Local Responses
Think like a seasoned analyst: Europe's Winter Energy Outlook: Local and Regional Risks
Tags: europe, eastern-europe, migration, border, local-analysis
Author: news_analytics_team
Created at: 2025-11-05 12:00:00
Average Rating:
Total Ratings:
Europe's Winter Energy Outlook: Local and Regional Risks
Analyze and think deeply about: Europe's Winter Energy Outlook: Local and Regional Risks
Tags: europe, energy, local, analysis
Author: news_analytics_team
Created at: 2025-11-05 12:00:00
Average Rating:
Total Ratings:

Curio AI Brain

Available in Chrome Web Store!