GoCoMA: Hyperbolic Multimodal Representation Fusion for Large Language Model-Generated Code Attribution
arXiv cs.CL / 4/21/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces GoCoMA, a multimodal framework to attribute generated code to the specific LLM source, addressing forensic needs created by LLMs producing human-like code.
- GoCoMA combines two complementary signals: code stylometry (structural and stylistic patterns) and image representations of binary pre-executable artifacts that capture lower-level, execution-oriented byte semantics.
- It embeds each modality in a hyperbolic Poincaré ball and uses a geodesic-cosine similarity-based cross-modal attention (GCSA) mechanism to fuse modalities effectively.
- The method maps the fused hyperbolic representation back to Euclidean space for final LLM-source attribution, then demonstrates consistent gains over unimodal and Euclidean multimodal baselines.
- Experiments on the CoDET-M4 and LLMAuthorBench open-source benchmarks show GoCoMA achieves stronger performance under the same evaluation protocols.
Related Articles

Every time a new model comes out, the old one is obsolete of course
Reddit r/LocalLLaMA

We built it during the NVIDIA DGX Spark Full-Stack AI Hackathon — and it ended up winning 1st place overall 🏆
Dev.to

Stop Losing Progress: Setting Up a Pro Jupyter Workflow in VS Code (No More Colab Timeouts!)
Dev.to

Building AgentOS: Why I’m Building the AWS Lambda for Insurance Claims
Dev.to

Where we are. In a year, everything has changed. Kimi - Minimax - Qwen - Gemma - GLM
Reddit r/LocalLLaMA