.. remoteconfigs: Remote configuration ==================== If you are using the base :meth:`Extractor <cognite.extractorutils.Extractor>` class, you can automatically fetch configuration from the extraction pipelines API in CDF. To use this, all you need to do is create a minimal configuration file like this: .. code-block:: yaml type: remote cognite: # Read these from environment variables host: ${COGNITE_BASE_URL} project: ${COGNITE_PROJECT} idp-authentication: token-url: ${COGNITE_TOKEN_URL} client-id: ${COGNITE_CLIENT_ID} secret: ${COGNITE_CLIENT_SECRET} scopes: - ${COGNITE_BASE_URL}/.default extraction-pipeline: external-id: my-extraction-pipeline containing only the ``cognite`` section, and the ``type: remote`` flag, and the configuration will be loaded dynamically from extraction pipelines. You can, for example, use the `upload config github action <https://github.com/cognitedata/upload-config-action>`_ to publish configuration files from a github repository. Remote configuration files are combined with local configuration and loaded into the extractor on startup. Detecting config changes ------------------------ When using the base class, you have the option to automatically detect new config revisions, and do one of several predefined actions (keep in mind that this is not exclusive to remote configs, if the extractor is running with a local configuration that changes, it will do the same action). You specify which with an ``reload_config_action`` enum. The enum can be one of the following values: * ``DO_NOTHING`` which is the default * ``REPLACE_ATTRIBUTE`` which will replace the ``config`` attribute on the object (keep in mind that if you are using the ``run_handle`` instead of subclassing, this will have no effect). Be also aware that anything that is set up based on the config (upload queues, cognite client objects, loggers, connections to source, etc) will not change in this case. * ``SHUTDOWN`` will set the ``cancellation_token`` event, and wait for the extractor to shut down. It is then intended that the service layer running the extractor (ie, windows services, systemd, docker, etc) will be configured to always restart the service if it shuts down. This is the recomended approach for reloading configs, as it is always guaranteed that everything will be re-initialized according to the new configuration. * ``CALLBACK`` is similar to ``REPLACE_ATTRIBUTE`` with one difference. After replacing the ``config`` attribute on the extractor object, it will call the ``reload_config_callback`` method, which you will have to override in your subclass. This method should then do any necessary cleanup or re-initialization needed for your particular extractor. To enable detection of config changes, set the ``reload_config_action`` argument to the ``Extractor`` constructor to your chosen action: .. code-block:: python # Using run handle: with Extractor( name="my_extractor", description="Short description of my extractor", config_class=MyConfig, version="1.2.3", run_handle=run_extractor, reload_config_action=ReloadConfigAction.SHUTDOWN, ) as extractor: extractor.run() # Using subclass: class MyExtractor(Extractor): def __init__(self): super().__init__( name="my_extractor", description="Short description of my extractor", config_class=MyConfig, version="1.2.3", reload_config_action=ReloadConfigAction.SHUTDOWN, ) The extractor will then periodically check if the config file has changed. The default interval is 5 minutes, you can change this by setting the ``reload_config_interval`` attribute. As with any other interval in extractor-utils, the unit is seconds.