LogoAIAny
  • Search
  • Collection
  • Category
  • Tag
  • Blog
LogoAIAny

Tag

Explore by tags

LogoAIAny

Curated AI Resources for Everyone

support@aiany.app
Product
  • Search
  • Collection
  • Category
  • Tag
Resources
  • Blog
Company
  • Privacy Policy
  • Terms of Service
  • Sitemap
Copyright © 2026 All Rights Reserved.
  • All

  • 30u30

  • ASR

  • ChatGPT

  • GNN

  • IDE

  • RAG

  • ai-agent

  • ai-api

  • ai-api-management

  • ai-client

  • ai-coding

  • ai-demos

  • ai-development

  • ai-framework

  • ai-image

  • ai-image-demos

  • ai-inference

  • ai-leaderboard

  • ai-library

  • ai-rank

  • ai-serving

  • ai-tools

  • ai-train

  • ai-video

  • ai-workflow

  • AIGC

  • alibaba

  • amazon

  • anthropic

  • audio

  • blog

  • book

  • bytedance

  • chatbot

  • chemistry

  • claude

  • claude-code

  • course

  • deepmind

  • deepseek

  • engineering

  • finance

  • foundation

  • foundation-model

  • gemini

  • github

  • google

  • gradient-booting

  • grok

  • huggingface

  • LLM

  • llm

  • math

  • mcp

  • mcp-client

  • mcp-server

  • meta-ai

  • microsoft

  • mlops

  • NLP

  • nvidia

  • ocr

  • ollama

  • openai

  • paper

  • physics

  • plugin

  • pytorch

  • RL

  • robotics

  • science

  • security

  • sora

  • translation

  • tutorial

  • vibe-coding

  • video

  • vision

  • xAI

  • xai

Icon for item

fairseq

2017
Facebook AI Research (FAIR)

fairseq is an open-source sequence modeling toolkit from Facebook AI Research (FAIR), implemented in Python on top of PyTorch. It provides reference implementations for a wide range of sequence models (Transformer, LSTM, Conv, wav2vec, wav2vec 2.0, etc.) and supports tasks such as machine translation, summarization, language modeling, and speech processing. Key features include multi-GPU and distributed training, fast generation (beam search, sampling, diverse beam), mixed-precision training, parameter/optimizer sharding, and many pre-trained models and examples. The project is MIT-licensed and documented at readthedocs.

githubNLPASRaudiotranslation+2
Icon for item

Buzz

2022
Chidi Williams (chidiwilliams)

Buzz is an open-source desktop and CLI tool that transcribes and translates audio offline using OpenAI's Whisper. It supports macOS, Windows and Linux, offers GUI and command-line interfaces, can be installed via native installers, Flatpak/Snap, winget or PyPI, and supports GPU acceleration via PyTorch/CUDA.

audioASRtranslationgithubpytorch+1
Icon for item

VideoCaptioner

2024
WEIFENG2333

VideoCaptioner is an AI-powered video subtitling assistant that combines ASR (local or cloud) with LLM-based subtitle segmentation, correction and translation. It supports offline GPU transcription, concurrent chunk transcription, VAD, speaker-aware processing, batch subtitling and one-click subtitle-to-video synthesis, with both GUI and CLI options.

videoai-videoaudioASRLLM+3
Icon for item

BabelDOC

2024
funstory-ai

BabelDOC is a PDF-focused document translation and bilingual comparison library/CLI that converts scientific papers and other PDFs (primarily English→Chinese) using OpenAI-compatible LLMs. It provides CLI and Python APIs, supports integration with self-hosted pipelines (e.g., PDFMathTranslate-next), offers online service through Immersive Translate, and includes many PDF-processing and translation options (OCR, glossary extraction, bilingual output, watermark controls).

ocrtranslationopenaiLLMai-tools+2

Neural Machine Translation by Jointly Learning to Align and Translate

2014
Dzmitry Bahdanau, Kyunghyun Cho +1

This paper introduces an attention-based encoder–decoder NMT architecture that learns soft alignments between source and target words while translating, eliminating the fixed-length bottleneck of earlier seq2seq models. The approach substantially improves BLEU, especially on long sentences, and matches phrase-based SMT on English-French without additional hand-engineered features. The attention mechanism it proposes became the foundation for virtually all subsequent NMT systems and inspired attention-centric models like the Transformer, reshaping machine translation and sequence modeling across NLP.

30u30paperNLPtranslation
  • Previous
  • 1
  • Next