Get started

To get started with Firebase Genkit, install the Genkit CLI and run genkit init in a Node.js project. The rest of this page shows you how.

Requirements

Node.js 20 or later.

Recommendation: The nvm and nvm-windows tools are a convenient way to install Node.

Install Genkit

Install the Genkit CLI by running the following command:

npm i -g genkit

This command installs the Genkit CLI into your Node installation directory so that it can be used outside of a Node project.

Create and explore a sample project

  1. Create a new Node project:

    mkdir genkit-intro && cd genkit-intro
    npm init -y

    Look at package.json and make sure the main field is set to lib/index.js.

  2. Initialize a Genkit project:

    genkit init
    1. Select your model:

      Gemini (Google AI)

      The simplest way to get started is with Google AI Gemini API. Make sure it's available in your region.

      Generate an API key for the Gemini API using Google AI Studio. Then, set the GOOGLE_GENAI_API_KEY environment variable to your key:

      export GOOGLE_GENAI_API_KEY=<your API key>

      Gemini (Vertex AI)

      If the Google AI Gemini API is not available in your region, consider using the Vertex AI API which also offers Gemini and other models. You will need to have a billing-enabled Google Cloud project, enable AI Platform API, and set some additional environment variables:

      gcloud services enable aiplatform.googleapis.com
      export GCLOUD_PROJECT=<your project ID>
      export GCLOUD_LOCATION=us-central1

      See https://cloud.google.com/vertex-ai/generative-ai/pricing for Vertex AI pricing.

    2. Choose default answers to the rest of the questions, which will initialize your project folder with some sample code.

    The genkit init command creates a sample source file, index.ts, which defines a single flow, menuSuggestionFlow, that prompts an LLM to suggest an item for a restaurant with a given theme.

    This file looks something like the following (the plugin configuration steps might look different if you selected Vertex AI):

    import * as z from 'zod';
    
    // Import the Genkit core libraries and plugins.
    import { generate } from '@genkit-ai/ai';
    import { configureGenkit } from '@genkit-ai/core';
    import { defineFlow, startFlowsServer } from '@genkit-ai/flow';
    import { googleAI } from '@genkit-ai/googleai';
    
    // Import models from the Google AI plugin. The Google AI API provides access to
    // several generative models. Here, we import Gemini 1.5 Flash.
    import { gemini15Flash } from '@genkit-ai/googleai';
    
    configureGenkit({
      plugins: [
        // Load the Google AI plugin. You can optionally specify your API key
        // by passing in a config object; if you don't, the Google AI plugin uses
        // the value from the GOOGLE_GENAI_API_KEY environment variable, which is
        // the recommended practice.
        googleAI(),
      ],
      // Log debug output to tbe console.
      logLevel: 'debug',
      // Perform OpenTelemetry instrumentation and enable trace collection.
      enableTracingAndMetrics: true,
    });
    
    // Define a simple flow that prompts an LLM to generate menu suggestions.
    export const menuSuggestionFlow = defineFlow(
      {
        name: 'menuSuggestionFlow',
        inputSchema: z.string(),
        outputSchema: z.string(),
      },
      async (subject) => {
        // Construct a request and send it to the model API.
        const llmResponse = await generate({
          prompt: `Suggest an item for the menu of a ${subject} themed restaurant`,
          model: gemini15Flash,
          config: {
            temperature: 1,
          },
        });
    
        // Handle the response from the model API. In this sample, we just convert
        // it to a string, but more complicated flows might coerce the response into
        // structured output or chain the response into another LLM call, etc.
        return llmResponse.text();
      }
    );
    
    // Start a flow server, which exposes your flows as HTTP endpoints. This call
    // must come last, after all of your plug-in configuration and flow definitions.
    // You can optionally specify a subset of flows to serve, and configure some
    // HTTP server options, but by default, the flow server serves all defined flows.
    startFlowsServer();
    

    As you build out your app's AI features with Genkit, you will likely create flows with multiple steps such as input preprocessing, more sophisticated prompt construction, integrating external information sources for retrieval-augmented generation (RAG), and more.

  3. Now you can run and explore Genkit features and the sample project locally on your machine. Download and start the Genkit Developer UI:

    genkit start

    Welcome to Genkit Developer UI

    The Genkit Developer UI is now running on your machine. When you run models or flows in the next step, your machine will perform the orchestration tasks needed to get the steps of your flow working together; calls to external services such as the Gemini API will continue to be made against live servers.

    Also, because you are in a dev environment, Genkit will store traces and flow state in local files.

  4. The Genkit Developer UI downloads and opens automatically when you run the genkit start command.

    The Developer UI lets you see which flows you have defined and models you configured, run them, and examine traces of previous runs. Try out some of these features:

    • On the Run tab, you will see a list of all of the flows that you have defined and any models that have been configured by plugins.

      Click menuSuggestionFlow and try running it with some input text (for example, "cat"). If all goes well, you'll be rewarded with a menu suggestion for a cat themed restaurant.

    • On the Inspect tab, you'll see a history of flow executions. For each flow, you can see the parameters that were passed to the flow and a trace of each step as they ran.

Next steps

Check out how to build and deploy your Genkit app with Firebase, Cloud Run, or any Node.js platform.