MCP Quick Start

Getting Started in 5 Minutes

Transform your data validation workflow with conversational AI in VS Code or Positron IDE. Here are three simple steps to start validating data through conversation (and no complex configuration required).

1. Install

pip install pointblank[mcp]

2. Configure Your IDE

For VS Code:

Option 1: Workspace Configuration (Recommended for teams)

  1. Create a .vscode/mcp.json file in your project folder
  2. Add this configuration:
{
  "servers": {
    "pointblank": {
      "command": "python",
      "args": ["-m", "pointblank_mcp_server.pointblank_server"]
    }
  }
}

Option 2: User Configuration (Personal use)

  1. Run command: MCP: Open User Configuration (Cmd/Ctrl + Shift + P)
  2. Add the same JSON configuration above

⚠️ Security Note: Only add MCP servers from trusted sources. VS Code will ask you to confirm trust when starting the server for the first time.

For Positron:

  1. Open Positron Settings
  2. Navigate to MCP Server configuration
  3. Add the configuration (format may vary)

Note: If you don’t see MCP settings, you may need to install an MCP extension first. Search for “MCP” in the Extensions marketplace.

3. Start Chatting

"Load my sales data and check its quality"

That’s basically how you get started.

Essential Commands

Master these five command patterns and you’ll be able to handle most data validation scenarios. Think of these as your fundamental vocabulary for talking to Pointblank.

Load Data

"Load the file /path/to/data.csv"
"Load my Netflix dataset from the working directory"
"Load the CSV file with sales metrics"
"Load customer_data.csv as my main dataset"

Explore Data

"Analyze the data for netflix_data"
"Show me a preview of the loaded data"
"Create a column summary table"
"Generate a missing values analysis"

What you’ll get: Comprehensive data profiling with statistics including missing values, data types, distributions, and summary statistics for each column. The preview and summary tables are automatically generated as beautiful HTML files that open in your browser. This gives you a complete picture of your dataset’s structure and characteristics before you define quality rules.

Check Quality

"Create a validator for netflix_data"
"Add validation that ratings are between 0 and 10"
"Check that all release years are reasonable"
"Apply the basic_quality template"

What you’ll get: Actual data quality validation that checks your data against business rules and domain knowledge. This tells you if your data meets your specific quality requirements and identifies rows that fail validation criteria.

Create Data Validations

"Add validation that show_id values are unique"
"Check that cast field is not empty for movies"
"Ensure vote_count is greater than 0"
"Validate that country field follows ISO format"

What you’ll get: Individual validation rules added to your validator. Each rule tests a specific business requirement and can be customized with thresholds and actions.

Run and Export

"Run the validation and show results"
"Export validation failures to CSV"
"Get failed rows for step 2"
"Save the validation report"

Top Three Use Cases

Once you have Pointblank set up, these scenarios show the most common and powerful ways to use natural language data validation. Each example demonstrates the conversational flow and immediate value you’ll experience.

1. 📊 Quick Data Check

When: You get a new dataset

👤 "I just got netflix_data.csv. Is it clean?"
🤖 "Loading dataset... Found 16,000 movies/shows with 18 columns.
    Quality analysis complete:
    - ✅ No duplicate show IDs
    - ⚠️ 466 missing countries (2.9%)
    - ⚠️ 132 missing directors (0.8%)
    Opening data preview in your browser..."

2. 🔄 Daily Validation

When: Regular data quality checks

👤 "Apply the basic_quality template to my data"
🤖 "Creating validator with basic quality checks...
    ✅ Data types validated
    ✅ Missing value thresholds checked
    ✅ Duplicate detection passed
    Validation complete! All checks passed."

3. 🚨 Issue Investigation

When: Something looks wrong

👤 "Show me movies with ratings above 9.5"
🤖 "Creating validator for high ratings...
    Found 47 entries with ratings > 9.5
    Exporting suspicious rows to CSV
    Most are documentaries - could be valid!"

These examples show how natural conversation can quickly identify and resolve data quality issues that might take hours to diagnose manually.

Core Capabilities

Pointblank’s MCP server provides powerful tools for comprehensive data validation with beautiful, interactive HTML reports:

Data Exploration

  • Interactive HTML previews with automatic browser opening showing head/tail rows
  • Column summary tables with detailed statistics and color-coded data types
  • Missing values analysis with visual patterns and percentages
  • Data quality analysis with comprehensive profiling insights

Validation Workflows

  • Validator creation with flexible thresholds and configuration
  • Many validation types for comprehensive data quality checking
  • Step-by-step validation building with natural language commands
  • Template-based validation for common data quality patterns

HTML Reports & Analysis

  • Interactive validation reports automatically opened in your browser
  • Timestamped HTML files for easy sharing and documentation
  • Python code generation for reproducible validation scripts

All interactions use natural language, making advanced data validation accessible to users at any technical level while producing publication-ready HTML reports.

Common Validation Rules

Understanding what validation rules to ask for will help you quickly build comprehensive data quality checks. These examples cover the most frequent validation scenarios using Pointblank’s built-in validation functions.

Data Integrity

  • “Check for duplicate show IDs”
  • “Ensure no missing required fields like title”
  • “Validate that release years are between 1900 and 2025”

Business Logic

  • “Ratings must be between 0 and 10”
  • “Budget must be positive numbers”
  • “Duration should be greater than 0”

Cross-Field Validation

  • “Release year should match date_added year”
  • “Vote count should correlate with popularity”
  • “Movies should have directors specified”

Available Templates

Pointblank includes pre-built validation templates:

  • basic_quality - Essential data quality checks
  • financial_data - Money and numeric validations
  • customer_data - Personal information validations
  • sensor_data - Time series and measurement checks
  • survey_data - Response and rating validations

These rule patterns can be combined and customized for your specific data and business requirements. The natural language interface makes it easy to express complex validation logic without learning technical syntax.

Some Tips and Tricks

These recommendations will help you get more value from your Pointblank MCP server and avoid some common pitfalls.

Talk Naturally

Good: “Check if customer emails look valid”

Avoid: “Execute col_vals_regex on email column”

Provide Context

Good: “This is for the board presentation”

Avoid: Just asking for validation without explanation

Build Incrementally

  1. Start with data profiling
  2. Add basic validation rules
  3. Create templates for reuse
  4. Set up automated checks

Save Templates

"Save these rules as 'customer_validation'"
"Apply the financial_data template"
"Use our standard survey validation"

Interactive Visual Tables

Pointblank automatically generates beautiful, interactive HTML tables for data exploration:

"Show me a preview of the data"
"Generate a column summary table"
"Create a missing values analysis"

These commands create professional HTML tables with:

  • Color-coded data types (numeric in purple, text in yellow)
  • Gradient styling tailored to each table type
  • Automatic browser opening for immediate viewing
  • Timestamped files for easy reference and sharing

The tables open automatically in your default browser, making it easy to share data insights with colleagues or include in presentations.

These practices help you build data quality workflows that scale with your needs while remaining accessible to those with varying technical backgrounds.

File Support

Pointblank works with many major data file formats, making it easy to validate data regardless of how it’s stored. This support means you can maintain consistent validation practices across your entire data ecosystem.

Type Extensions Example Backend Support
CSV .csv sales_data.csv pandas, polars
Parquet .parquet big_data.parquet pandas, polars
JSON .json api_response.json pandas, polars
JSON Lines .jsonl streaming_data.jsonl pandas, polars

The consistent natural language interface works the same regardless of file format, so you can focus on validation logic rather than technical details. Polars provides faster processing for large datasets, while Pandas offers broader format support.

Quick Troubleshooting

When you encounter issues, these quick fixes resolve the most common problems. Furthermore, the natural language interface means you can always ask for help and explanations.

Problem Quick Fix
“File not found” Use absolute path: /Users/name/Downloads/data.csv
“DataFrame not found” Check loaded datasets with “List my loaded dataframes”
“Validator not found” Use “List active validators” to see available validators
“Validation too slow” Try “Use pandas backend” or sample your data first
“HTML tables won’t open” Check your default browser settings
“Need validation ideas” Ask “Show me validation templates” or “Suggest validations for my data”

Browser Issues: The HTML tables automatically open in your default browser. If they don’t appear, check that your browser isn’t blocking pop-ups and that you have a default browser set in your system preferences.

Remember, you can always ask the AI to explain what’s happening or suggest solutions when you run into problems.

Now You’re Ready!

You now have everything needed to start validating data through conversation. The beauty of Pointblank’s MCP server is that it grows with your expertise: start simple and gradually build more sophisticated validation workflows as you become comfortable with the interface.

Start with simple commands and build up to more complex validation workflows. The AI will guide you through the process and help you create robust data quality checks!