Explore >> Select a destination


You are here

leogao.dev
| | blog.research.google
5.2 parsecs away

Travel
| | [AI summary] Google Research introduces TimesFM, a decoder-only foundation model for time-series forecasting with zero-shot capabilities, pre-trained on 100 billion real-world time-points, outperforming existing methods in various domains.
| | research.google
5.0 parsecs away

Travel
| | Posted by Adam Roberts, Staff Software Engineer and Colin Raffel, Senior Research Scientist, Google Research Over the past few years, transfer le...
| | slingingbits.com
6.0 parsecs away

Travel
| | After benchmarking the R1 1776 model and seeing how post training influenced its performance (full post here), I realized another gap. Models that can technically handle a huge context window often degrade long before you hit their max token limit. Plenty of benchmarks test for raw throughput or model quality
| | ljvmiranda921.github.io
16.0 parsecs away

Travel
| In this blog post, I want to demonstrate how we can leverage large language models like GPT-3 as a viable affordance to reduce a human annotator's cognitive ...