Skip to content

User Guide Overview

Welcome to Citation Compass! This guide helps you choose the right interface for your needs and get started with common workflows.


Platform Access

Choose your interface based on how you work:

Interface When to Use Quick Start
πŸ–₯️ Dashboard Interactive exploration, demos, quick analysis streamlit run app.py
πŸ““ Notebooks Reproducible research, custom analysis, model training jupyter notebook notebooks/
πŸ”Œ Python API Automation, integration with existing tools from src.services import get_ml_service

New to Citation Compass? Start with the dashboardβ€”it requires no code and provides immediate visualization.


User Personas

Different users have different needs. We've designed Citation Compass to support three main personas, each with tailored workflows and entry points.

User Journey

πŸŽ“ Academic Researchers

You want: Citation analysis, research discovery, network visualization

Start here:

  1. Interactive Features - Dashboard walkthrough

  2. Network Analysis - Community detection and metrics

  3. Results Interpretation - Export for publications

πŸ€– Data Scientists

You want: Custom models, batch predictions, reproducible pipelines

Start here:

  1. Notebook Pipeline - 4-notebook workflow

  2. ML Predictions - Model training and evaluation

  3. Developer Guide - API customization

πŸ“Š Research Administrators

You want: Usage monitoring, team reports, performance tracking

Start here:

  1. Interactive Features - Dashboard overview

  2. Results Interpretation - Report generation

  3. Network Analysis - Performance metrics


The user journey diagram shows the typical flow from initial exploration through data analysis to final publication. Notice how most users start with demo mode to learn features before importing their own data.


Getting Started Paths

Goal: Try features with demo data (no setup required)

Steps: 1. Launch: streamlit run app.py 2. Load demo dataset from sidebar 3. Try Interactive Features 4. Export results

Time: 10 minutes

Goal: Analyze your citation network comprehensively

Steps: 1. Configure Neo4j database (setup guide) 2. Import data via Data Import 3. Follow Notebook Pipeline 4. Train models with ML Predictions

Time: 2-3 hours

Goal: Integrate into existing research workflows

Steps: 1. Review Developer Guide 2. Check API Reference 3. Use Python API for automation 4. Build custom dashboards/reports

Time: Varies by complexity


Common Workflows

πŸ” Citation Discovery

Goal: Find related papers you might have missed

flowchart LR
    A[Input Paper] --> B[Generate Predictions]
    B --> C[Explore Embeddings]
    C --> D[Validate Results]
    D --> E[Export Reading List]

    style A fill:#e3f2fd
    style B fill:#fff3e0
    style C fill:#e8f5e8
    style D fill:#fce4ec
    style E fill:#f1f8e9
  1. Input: Choose a paper from your network
  2. Predict: Run ML predictions (guide)
  3. Explore: Visualize in embedding space
  4. Validate: Check against known citations
  5. Export: Save reading list with scores

πŸ•ΈοΈ Network Analysis

Goal: Understand citation communities and influence

flowchart LR
    A[Load Network] --> B[Compute Metrics]
    B --> C[Detect Communities]
    C --> D[Analyze Trends]
    D --> E[Generate Report]

    style A fill:#ffebee
    style B fill:#e0f2f1
    style C fill:#f3e5f5
    style D fill:#e8f5e8
    style E fill:#fff3e0
  1. Load: Import data or use demo
  2. Metrics: Calculate centrality (guide)
  3. Communities: Run Louvain or label propagation
  4. Trends: Analyze temporal patterns
  5. Report: Export LaTeX tables

πŸ€– Model Training

Goal: Train custom ML model on your data

flowchart LR
    A[Prepare Data] --> B[Train Model]
    B --> C[Evaluate]
    C --> D[Predict]
    D --> E[Deploy]

    style A fill:#f1f8e9
    style B fill:#e3f2fd
    style C fill:#fce4ec
    style D fill:#fff3e0
    style E fill:#e8f5e8
  1. Data: Import citation network
  2. Train: Use notebook 02 for TransE training
  3. Evaluate: Check MRR, Hits@K metrics
  4. Predict: Generate citations
  5. Deploy: Save model for dashboard use

Feature Matrix

Feature Availability by Interface (click to expand)
Feature Dashboard Notebooks API
Citation Prediction βœ… Interactive βœ… Customizable βœ… Programmatic
Network Analysis βœ… Visual βœ… Detailed βœ… Batch
Community Detection βœ… Real-time βœ… Multiple Algorithms βœ… Scalable
Temporal Analysis βœ… Interactive βœ… Advanced βœ… Automated
Export Capabilities βœ… Basic βœ… Advanced βœ… Custom
Model Training ❌ βœ… Full Pipeline βœ… Programmatic
Custom Visualization ❌ βœ… Matplotlib/Plotly βœ… Programmatic
Batch Processing ❌ βœ… Yes βœ… Scalable

Dashboard: Best for exploration and demos Notebooks: Best for research and custom analysis API: Best for automation and integration


Best Practices

Analysis Tips

  • Start small: Use demo datasets before loading large networks
  • Validate results: Cross-check predictions with domain expertise
  • Document settings: Record parameters for reproducibility
  • Iterate: Dashboard for exploration β†’ notebooks for final analysis

Technical Tips

  • Monitor resources: Track memory usage with large datasets
  • Enable caching: Speeds up repeated analyses significantly
  • Check logs: Look in logs/ directory for debugging
  • Version control: Track notebooks and config files

Reporting Tips

  • Document methodology: Explain analysis approach clearly
  • Consistent styling: Use same color schemes across visualizations
  • Include metrics: Add confidence intervals and statistical tests
  • High resolution: Export figures at 300+ DPI for publications

Support & Community

Documentation:

Get Help:


Next Steps

Choose your path forward:


Happy analyzing! πŸ”¬βœ¨