ANLAM-Net Lab · Research Programme
Six interconnected research strands spanning language models, multimedia intelligence, and evaluation science.
We advance the adaptability, robustness, and efficiency of large language models across diverse domains and languages. Our work spans parameter-efficient fine-tuning (LoRA, QLoRA, adapters), cross-lingual transfer, instruction-following, alignment, and comprehensive evaluation, with strong focus on Turkish and morphologically complex languages.
Joint modeling of text, image, and audio for aspect-based sentiment prediction — fusing cross-modal attention with aspect-specific pooling mechanisms.
Construction of a large-scale, aspect-annotated Turkish social media dataset for benchmarking NLP systems on morphologically rich text.
Hierarchical encoder architectures that aggregate information across dozens of source documents while maintaining factual grounding and attribution.
Efficient attention mechanisms and memory-augmented transformers for processing 10k+ token inputs in legal and scientific document summarization.
A multi-task evaluation framework for Turkish covering QA, NER, NLI, summarization, and generation with rigorous human validation protocols.
Systematic stress-testing of NLP models using character-level attacks, paraphrase-based perturbations, and domain-shift evaluation scenarios.
Fusion of speech prosody, facial expression, and text for fine-grained multimodal sentiment prediction in video content.
Contrastive learning approaches for aligning text and visual representations, enabling zero-shot retrieval across modalities.
Dense retrieval systems coupled with generative models for knowledge-grounded text synthesis in domain-specific and multilingual settings.
Taxonomy and mitigation strategies for factual hallucinations in neural text generation, including post-hoc verification and constrained decoding.
All claims are backed by controlled experiments with statistical significance testing, ablation studies, and reproducible baselines.
Datasets, model checkpoints, and evaluation code released openly to accelerate community progress in multilingual and multimedia NLP.
Researchers from four universities contribute complementary expertise in linguistics, engineering, AI systems, and applied ML.
Google Cloud Research Credits enable large-scale model training and comprehensive evaluation sweeps across multilingual benchmarks.
We prioritize underrepresented languages — especially Turkish and Turkic families — developing resources for communities with limited NLP tooling.
Safety, fairness, and interpretability are embedded from research design through evaluation, with explicit bias audits on all released systems.