
Optimizing Sequence Models for Dynamical Systems
Ablation study deconstructing sequence models. Attention-augmented Recurrent Highway Networks outperform Transformers on …...

Ablation study deconstructing sequence models. Attention-augmented Recurrent Highway Networks outperform Transformers on …...

Enhanced TabMe benchmark for page stream segmentation, creating TabMe++, showing fine-tuned decoder-based LLMs …...
GTR-CoT uses graph traversal chain-of-thought reasoning to improve optical chemical structure recognition....
αExtractor uses ResNet-Transformer to extract chemical structures from literature images, including noisy and hand-drawn …...
Chen et al.'s dual-stream encoder approach for robust molecular structure recognition from diverse real-world images …...
MolParser converts molecular images from scientific documents to machine-readable formats using E-SMILES....

Lu et al. introduce SpaceFormer, a Transformer that models entire 3D molecular space (not just atoms) for superior …...

LLM applications for insurance document automation using parameter-efficient fine-tuning and analysis of calibration …...
Skinnider (2024) shows that generating invalid SMILES actually improves chemical language model performance through …...

Learn how dataset bias can lead to misleading results in NLP: a sarcasm detection model that actually learned to …
Campos & Ji's method for converting 2D molecular images to SMILES strings using Transformers and SELFIES representation....

Undergraduate thesis exploring representation learning for social media text and developing tools for cross-platform …