Nexconn-AI
    • Nexconn LLM API Service
    • Quick Start
    • Key Concepts and Terminology
    • Try Nexconn Large Model API Online
    • How to Get API-KEY
    • FAQ
    • Model
      • Model List
        GET
    • Chat
      • Bypass Interface
        • Bypass Anthropic Protocol
        • bypass Vertex/Gemini protocol
        • bypass Responses protocol
      • gemini-3.1-pro-preview
        • gemini-3.1-pro-preview reasoning
        • gemini-3.1-pro-preview qfile
      • gemini-2.5-pro
        • gemini-2.5-pro reasoning
      • openai/gpt-5
        • openai/gpt-5 reasoning
      • openai/gpt-5.2
        • openai/gpt-5.2 Reasoning
      • openai/gpt-5.2-codex
        • openai/gpt-5.2-codex
      • gemini-2.5-flash
        • gemini-2.5-flash Reasoning
      • claude-4.5-sonnet
        • claude-4.5-sonnet reasoning
      • claude-4.6-opus
        • claude-4.6-opus reasoning
      • deepseek/deepseek-v3.2-251201
        • deepseek/deepseek-v3.2-251201 reasoning
      • deepseek-v3
        • Chat
      • claude-3.7-sonnet
        • Chat completion
      • doubao-seed-1.6
        • doubao-seed-1.6
      • qwen3-max-2026-01-23
        • Chat completion
      • moonshotai/kimi-k2.5
        • Chat completion
      • Chat completion
        POST
      • Anthropic Protocol
        POST
    • Video
      • Veo
        • Create Video Generation Task
        • Query Video Generation Task
      • sora-2
        • Create Video Generation Task
        • Query Video Generation Status
        • Video Remix
      • sora-2-pro
        • Create Video Generation Task
        • Query Video Generation Status
        • Video Remix
      • kling-v2-1
        • Create Video Task
        • Query Video Generation Status
      • kling-v2-5-turbo
        • Create Video Task
        • Query Video Generation Status
      • kling-v2-6
        • Create Video Task
        • Query Video Generation Status
      • kling-v3
        • Create Video Task
        • Query Video Generation Status
      • kling-video-o1
        • Create Video Task
        • Query Video Generation Status
      • kling-v3-omni
        • Create Video Task
        • Query Video Generation Status
      • bytedance/doubao-seedance-2-0-260128
        • Create Video Generation Task
        • Query Video Generation Task
      • vidu
        • viduq1
          • Create text-to-video task
          • Create reference to video task - Non-subject invocation (video generation)
          • Create reference to video task - Subject invocation (supports video with audio function)
          • Query task status
          • Query task results
        • viduq2
          • Create text-to-video task
          • Create reference to video task - Non-subject invocation (video generation
          • Create reference to video task - Subject invocation (supports video with audio function)
          • Query task status
          • Query task results
        • viduq2-pro
          • Create Image-to-Video Task
          • Create First and Last Frame to Video Task
          • Create reference video generation task - non-subject invocation (video generation)
          • Query task status
          • Query task results
        • viduq2-turbo
          • Create Image-to-Video Task
          • Create First and Last Frame to Video Task
          • Query task status
          • Query task results
        • viduq3-pro
          • Create Image-to-Video Task
          • Create text-to-video task
          • Create First and Last Frame to Video Task
          • Query task status
          • Query task results
        • viduq3-turbo
          • Create Image-to-Video Task
          • Create text-to-video task
          • Create First and Last Frame to Video Task
          • Query task status
          • Query task results
    • Image Generation
      • kling-v1
        • Create text-to-image or single image-to-image task
        • Query task status
      • kling-v1-5
        • Create text-to-image or single image-to-image task
        • Query task status
      • kling-v2
        • Create text-to-image or single image-to-image task
        • Create multi-image-to-image task
        • Query task status
      • kling-v2-new
        • Create single image-to-image task
        • Query task status
      • kling-v2-1
        • Create text-to-image or single image-to-image task
        • Create multi-image-to-image task
        • Query task status
      • gemini-2.5-flash-image
        • Chat interface - supports text-to-image, image-to-image, and pure conversation
        • Text-to-image API - Generate images from text descriptions
        • Image-to-image API - Generate new images based on input images
      • gemini-3.0-pro-image-preview
        • Chat interface - supports text-to-image, image-to-image, and pure conversation
        • Text-to-image API - Generate images from text descriptions
        • Image-to-image API - Generate new images based on input images
      • gemini-3.1-flash-image-preview
        • Chat interface - supports text-to-image, image-to-image, and pure conversation
        • Text-to-image API - Generate images from text descriptions
        • Image-to-image API - Generate new images based on input images
      • kling-image-o1
        • Create image generation task
        • Query image generation task
        • Get Result
    • Files
      • Create file upload task
      • Query file status
      • List user files
    • Schemas
      • Chat
        • ChatCompletionRequest
        • ChatCompletionRequestMessage
        • ToolObject
        • ChatTemplateKwargs
        • ThinkType
        • ReasoningType
        • ImageConfig
        • SafetySetting
        • MessageContent
        • ImageUrl
        • VideoUrl
        • InputAudio
        • FileUrl
        • CacheControl
        • FunctionCall
        • ToolCall
        • ToolCallFunction
        • Image
        • ThinkingBlock
        • ToolFunction
        • ToolParameters
      • Video
        • kling-v2-1
          • KlingV21CreateRequest
          • KlingV21CreateResponse
          • KlingV21StatusResponse
        • kling-v2-5-turbo
          • KlingV25TurboCreateRequest
          • KlingV25TurboCreateResponse
          • KlingV25TurboStatusResponse
        • kling-video-o1
          • KlingVideoO1CreateRequest
          • KlingVideoO1CreateResponse
          • KlingVideoO1StatusResponse
        • kling-v3-omni
          • KlingV3OmniCreateRequest
        • kling-v3
          • KlingV3CreateRequest
        • kling-v2-6
          • KlingV26CreateRequest
          • KlingV26CreateResponse
          • KlingV26VideoStatusResponse
        • Veo
          • CreateVideoGenerationRequest
          • CreateVideoGenerationResponse
          • VideoGenerationJobInfo
          • ErrorResponse
          • Instance
          • ImageInput
          • LastFrameInput
          • VideoInput
          • ReferenceImage
          • Parameters
          • VideoGenerationData
          • VideoResult
        • VideoCreateResponse
      • Image Generation
        • kling-v1
          • KlingV1CreateImageRequest
        • kling-v1-5
          • KlingV15CreateImageRequest
        • kling-v2
          • KlingV2CreateImageRequest
          • KlingV2EditImageRequest
        • kling-v2-new
          • KlingV2NewCreateImageRequest
        • kling-v2-1
          • KlingV21CreateImageRequest
          • KlingV21EditImageRequest
        • gemini-2.5-flash-image
          • Gemini25FlashImageChatCompletionRequest
          • Gemini25FlashImageGenerationRequest
          • Gemini25FlashImageEditRequest
          • Gemini25FlashImageConfig
        • gemini-3.0-pro-image-preview
          • Gemini30ProImageChatCompletionRequest
          • Gemini30ProImageGenerationRequest
          • Gemini30ProImageEditRequest
          • Gemini30ProImageConfig
        • KlingImageTaskResponse
        • KlingImageTaskStatusResponse
        • ChatCompletionResponse
        • ImageGenerationResponse
        • FalOmniImageRequest
        • Gemini31FlashImageChatCompletionRequest
        • Gemini31FlashImageGenerationRequest
        • Gemini31FlashImageEditRequest
        • ChatMessage
        • Gemini31FlashImageConfig
      • CreateFileRequest
      • QueueStatus
      • ContentItem
      • FileResponse
      • Veo31FirstLastFrameToVideoInput
      • ToolItem
      • CreateVideoTaskResponse
      • FileListResponse
      • ApiErrorBody
      • FileDeleteResponse
      • GetVideoTaskResponse
      • Veo31ImageToVideoInput
      • Veo31ImageToVideoOutput
      • ChatCompletionResponse
      • FileError
      • ImageGenerationResponse
      • ErrorResponse
      • KodoSource
      • File
      • ImageUrlObject
      • VideoUrlObject
      • AudioUrlObject
      • DraftTaskObject
      • ChatMessage
      • VideoTaskOutputContent
      • VideoTaskToolUsageItem
      • VideoTaskUsage

    Key Concepts and Terminology

    ๐Ÿ“š

    Key Concepts and Terminology#

    Helps you quickly understand professional terms related to AI large models
    TIP
    This document uses plain language to explain various technical terms, making them easy to understand even without a technical background.

    Core Concepts#

    ๐Ÿค– Large Language Model (LLM)#

    ๐Ÿ’ก
    Plain Explanation: A large model is like a "super-intelligent AI brain" that can understand and generate human language by learning from massive amounts of text data, completing various intelligent tasks.
    Characteristics of Large Models:
    โœ“ Massive Scale: Contains billions or even trillions of parameters (think of them as "knowledge points")
    โœ“ Comprehensive Capabilities: Can handle text, code, translation, Q&A, and many other tasks
    โœ“ Continuous Learning: Constantly improves intelligence through training
    Common Large Models: Kimi, Tongyi Qianwen, Wenxin Yiyan, etc.

    ๐Ÿ”ค Token#

    ๐Ÿ“Š
    Plain Explanation: A token is the smallest unit for AI to process text, similar to a "measurement unit" for text.

    Token Calculation#

    LanguageToken CalculationExample
    ChineseUsually 1-2 characters = 1 token"Hello world" โ‰ˆ 4 tokens
    EnglishUsually 1 word = 1-2 tokens"Hello World" โ‰ˆ 2 tokens
    CodeCalculated based on characters and structureprint("Hello") โ‰ˆ 4-5 tokens
    INFO
    Why understand tokens?
    Because AI service billing is typically based on token usage. The longer the input and output text, the more tokens consumed, and the higher the cost.
    Real Example: If you ask the AI "How is the weather today?" (about 8 tokens), and the AI responds "The weather is sunny today, the temperature is 25 degrees, and it is suitable for going out." (about 20 tokens), this conversation consumes approximately 28 tokens in total.

    ๐Ÿ”— API (Application Programming Interface)#

    ๐Ÿ”Œ
    Plain Explanation: An API is a "bridge" for software communication, allowing your application to call Nexconn's AI services.

    Analogy#

    ๐Ÿช API is like a restaurant "menu"
    You (the application) order from the menu (API) to the kitchen (AI service), then receive your dish (AI response).

    Practical Purpose#

    Through APIs, you can integrate AI capabilities into your website, app, or software without developing complex AI systems yourself.
    API Types Supported by Nexconn:
    OpenAI API: Compatible with OpenAI interface standards
    Anthropic API: Compatible with Anthropic (Claude) interface standards
    Standard REST API: Universal network interface protocol

    ๐Ÿง  Inference#

    ๐Ÿ”ฎ
    Plain Explanation: Inference is the process where an AI model "thinks" through input questions or instructions and provides answers.

    Inference Process Example#

    Question: "Help me write a thank-you letter"
    1.
    Input Stage โ†’ User submits question
    2.
    Inference Stage โ†’ AI analyzes request and generates content
    3.
    Output Stage โ†’ Generates complete thank-you letter text
    NOTE
    Inference Service provides this AI "thinking" and "answering" capability as a cloud service.

    ๐ŸŽฏ Prompt#

    โœ
    Plain Explanation: A prompt is the "instruction" or "question" you give to the AIโ€”the content you input.
    Prompt quality directly affects AI response quality!

    Comparison Examples#

    TypeExampleEvaluation
    โŒ Poor Prompt"Write something"Too vague, AI doesn't know what to write
    โœ… Good Prompt"Please help me write a 300-word product introduction for a smartwatch, highlighting health monitoring features"Clear and specific, AI can accurately understand the requirement
    TIP
    Tips for Writing Good Prompts:
    Clearly define the task objective
    Provide sufficient background information
    Specify output format or style
    Give specific constraints

    Advanced Features#

    ๐Ÿค– MCP (Model Context Protocol)#

    ๐Ÿ”ง
    Plain Explanation: MCP is a technology that gives AI "superpowers" by adding various tool capabilities.

    Analogy#

    If you think of AI as a smart assistant, then MCP equips this assistant with various "toolboxes":
    ๐Ÿ” Search Toolbox: Enables AI to search the web for latest information
    ๐Ÿ“Š Data Toolbox: Enables AI to analyze and process data
    ๐Ÿ“ File Toolbox: Enables AI to read and edit files
    ๐Ÿ“ง Communication Toolbox: Enables AI to send emails or messages
    Practical Use Cases:
    AI can retrieve real-time stock prices (using financial data tools)
    AI can help you book restaurants (using reservation system tools)
    AI can generate and save documents (using file processing tools)

    ๐Ÿ”ง Agent#

    ๐Ÿคน
    Plain Explanation: An agent is an "AI assistant" that can autonomously complete complex tasksโ€”it not only answers questions but also proactively uses tools to solve problems.

    Regular AI vs Agent Comparison#

    ComparisonRegular AIAgent
    Working MethodPassively answers questionsโœ… Proactively executes tasks
    Capability RangeMainly text generationโœ… Can use multiple tools
    Task ComplexitySingle-step simple tasksโœ… Multi-step complex tasks
    AutonomyRequires explicit instructionsโœ… Can autonomously plan steps

    Agent Work Example#

    Task: "Help me organize last week's sales data and send a report to my boss"
    Execution Steps:
    1.
    Read last week's sales data from database
    2.
    Analyze data and generate charts
    3.
    Write analysis report
    4.
    Send via email to boss
    5.
    โœ… Task complete!

    ๐ŸŒ Web Search#

    ๐Ÿ”
    Plain Explanation: Enables AI to access the internet and retrieve the latest real-time information.

    Comparison Examples#

    โŒ Without Web Search
    Q: "What's the weather in Beijing today?"
    A: "Sorry, I cannot retrieve real-time weather information..."
    โœ… With Web Search
    Q: "What's the weather in Beijing today?"
    A: "Beijing is sunny today, temperature 15-25โ„ƒ, air quality is good."
    Use Cases:
    Query real-time information (weather, news, stocks, etc.)
    Get latest materials (latest technical documentation, policies and regulations, etc.)
    Verify factual information (check data, confirm message authenticity, etc.)

    ๐Ÿ’ญ Chain of Thought (Reasoning)#

    ๐Ÿงฉ
    Plain Explanation: Enables AI to "think step by step" like humans, rather than directly giving answers, thereby improving response quality for complex problems.

    Comparison Example#

    Question: "Xiao Ming has 15 apples, ate 3, bought 8 more, gave 5 to friendsโ€”how many does he have left?"
    โŒ Regular Mode
    "Xiao Ming has 15 apples left." (Direct answer, prone to errors)
    โœ… Chain of Thought Mode
    Let me calculate step by step:
    1.
    Initial: 15 apples
    2.
    Ate 3: 15 - 3 = 12
    3.
    Bought 8: 12 + 8 = 20
    4.
    Gave 5 to friends: 20 - 5 = 15
    5.
    Final answer: 15 apples
    Advantages:
    Reduces calculation errors
    More rigorous logical reasoning
    More transparent answer process

    ๐Ÿ’ฌ Context Memory#

    ๐Ÿง 
    Plain Explanation: AI can "remember" previous conversation content, enabling coherent multi-turn conversations.

    Conversation Example#

    Round 1
    ๐Ÿ‘ค You: "I want to travel to Japan"
    ๐Ÿค– AI: "Japan is a beautiful country! When are you planning to go?"
    Round 2
    ๐Ÿ‘ค You: "Next spring"
    ๐Ÿค– AI: "Spring is perfect for visiting Japanโ€”you can enjoy cherry blossoms! I recommend Tokyo, Kyoto, and Osaka."
    Round 3
    ๐Ÿ‘ค You: "What do I need to prepare?"
    ๐Ÿค– AI: "For traveling to Japan, you need to prepare: 1. Passport and visa 2. Clothing suitable for spring 3. Japanese yen cash 4. Translation software..."
    ๐Ÿ’ก Note: The AI remembered the context of "going to Japan" and "next spring"
    Application Value:
    More natural conversation experience
    No need to repeat background information
    Supports complex multi-turn interactions

    Billing Terms#

    ๐Ÿ’ฐ Pay-As-You-Go#

    Meaning: Charged based on actual resource usage (token count)โ€”pay for what you use.
    Analogy: Like utility billsโ€”you pay for the electricity you actually use; no usage means no charge.
    Advantages:
    โœ“ Controllable costs, suitable for scenarios with unstable usage
    โœ“ No prepayment required, lowers barrier to entry
    โœ“ Flexible scaling, use as needed

    ๐Ÿ“ฆ Resource Package#

    Meaning: Purchase a certain amount of token quota in advance and enjoy discounted rates.
    Analogy: Like buying an annual gym membershipโ€”purchasing at once is more cost-effective than paying each time.
    Suitable For:
    โœ“ Users with high usage frequency
    โœ“ Users seeking cost-effectiveness
    โœ“ Enterprises with stable usage needs

    ๐Ÿ”€ Unified Token#

    Meaning: A universal token quota that can be shared across multiple different AI models. Nexconn's AI large model unified resource package token count is based on DeepSeek-V3.1 as the baseline. When using other models, the token deduction will be proportionally adjusted based on different prices (i.e., more expensive models have higher deduction multiples, output context longer than input has higher deduction multiples, cheaper models have lower deduction coefficients). Through a unified deduction standard, you can freely switch between AI models at different price points and enjoy a convenient billing experience. Please log in to the Nexconn Large Model API Console to check token consumption and request records in real time.
    Analogy: Like recharging a "universal shopping card" that can be used at multiple stores, not limited to just one.
    Advantages:
    โœ“ Flexibly switch between different models
    โœ“ Avoid wasting quota on a single model
    โœ“ Simplify account management

    Technical Architecture Terms#

    ๐Ÿ” End-to-End Encryption#

    ๐Ÿ”’
    Plain Explanation: Data is encrypted throughout transmissionโ€”only the sender and receiver can see the content; no one in between can intercept it.
    Protected Content:
    API keys
    User input questions
    AI returned answers
    All sensitive business data

    ๐Ÿ“ก REST API#

    ๐ŸŒ
    Plain Explanation: REST API is a web communication standard that enables different software systems to interact over the internet.

    How It Works (Simplified)#

    1.
    Your application sends a request โ†’ "Please help me generate an article"
    2.
    Nexconn AI service receives and processes the request
    3.
    AI returns a response โ†’ Generated article content
    4.
    Your application displays the result to the user
    Characteristics:
    Standardized: Follows unified specifications
    Simple to Use: Easy to understand and implement
    Cross-Platform: Supports various programming languages

    ๐Ÿ“š SDK (Software Development Kit)#

    ๐Ÿ› 
    Plain Explanation: An SDK is an official "toolbox" containing ready-made code and tools that make it easier for developers to use AI services.

    Comparison Example#

    โŒ Without SDK
    Need to write extensive code for network request encapsulation
    Need to handle data format conversion
    Need to implement error handling logic
    Need to complete authentication and authorization flow
    โฑ๏ธ Long development time, prone to errors
    โœ… With SDK
    Just a few lines of code:
    โšก Fast development, stable and reliable
    Supported Languages: Python, Java, JavaScript, Go, etc.

    Model-Related Terms#

    ๐ŸŽ›๏ธ Parameters#

    ๐Ÿ”ข
    Plain Explanation: Parameters are the "knowledge units" that make up an AI modelโ€”more parameters usually means a "smarter" model.
    Model SizeParameter CountCapability Characteristics
    Small ModelHundreds of millionsFast response, low cost, suitable for simple tasks
    Medium ModelTens of billionsBalances performance and cost
    Large ModelHundreds of billions ~ trillionsMost capable, suitable for complex tasks
    NOTE
    Models with more parameters typically have higher computational costs and correspondingly higher API call fees.

    ๐Ÿ”„ Fine-tuning#

    ๐ŸŽฏ
    Plain Explanation: Based on a general large model, conduct specialized training for specific domains or tasks to make it better at handling specific problems.

    Analogy#

    Like a general practitioner (general model) who, after specialized training, becomes a cardiologist (fine-tuned model), more professional in the field of heart disease.
    Use Cases:
    ๐Ÿ“ Enterprise Customer Service: Train AI to specifically answer company product questions
    โš–๏ธ Legal Assistant: Train AI proficient in legal provisions
    ๐Ÿฅ Medical Advisor: Train AI familiar with medical knowledge

    ๐ŸŒก๏ธ Temperature#

    ๐ŸŽฒ
    Plain Explanation: The temperature parameter controls the "creativity level" or "randomness" of AI responses.

    Temperature Comparison#

    Temperature RangeCharacteristicsSuitable Scenarios
    Low Temperature (0-0.3) ๐ŸงŠStable, conservative, accurate responsesData analysis, code generation
    Medium Temperature (0.5-0.7) ๐ŸŒค๏ธBalanced, natural responsesDaily conversation, Q&A
    High Temperature (0.8-1.0) ๐Ÿ”ฅCreative, diverse, random responsesCreative writing, brainstorming
    Modified atย 2026-04-24 10:42:45
    Previous
    Quick Start
    Next
    Try Nexconn Large Model API Online
    Built with