Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Jupyter Telemetry Enhancement Proposal #41

Closed
wants to merge 5 commits into from
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
13 changes: 13 additions & 0 deletions 30-telemetry/press_release.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,13 @@
# Jupyter releases a new Telemetry System

Today the Jupyter Team announced the release of its newest feature—the Jupyter Telemetry System. This new system enables anyone deploying Jupyter products to safely collect structured “event” data from Jupyter applications. This makes it easier to secure Jupyter deployments, detect intrusion, improve user experience, scale resources effectively, and minimize costs for both administrators and users. It achieves all of this while prioritizing user privacy. The result is an improved Jupyter experience for all.

Jupyter’s telemetry system has been a long time coming. Members of the Jupyter community have been exploring various ways to collect structured events inside Jupyter applications. It’s no easy task. While event data can improve the development and deployment of Jupyter applications, it’s difficult to create an effective, safe and secure event-logging system.

This new telemetry system strikes a successful balance between effective event-logging and honoring user privacy. Transparency is a key tenant. Users are always informed when data is being collected. In most cases, users can “opt-in” and dictate which events are logged (aligning with GDPR). In stricter Jupyter environments where users cannot dictate the event-logging (e.g. FEDRAMP, FERPA, HIPAA-compliant services), users will be informed by clear and transparent messages from Jupyter’s telemetry system.

While staying user-friendly, the telemetry system is both flexible and extensible. It is highly configurable and integrates seamlessly with Jupyter-based applications and can be extended to work with non-Jupyter applications. Jupyter extension developers can leverage its well-defined APIs and libraries to publish events to various storage systems. Jupyter administrators can dictate what events are collected and when. It can store code executed on Jupyter Kernels, track clicks by users inside JupyterLab, record which datasets have been opened. For more sensitive or secure Jupyter deployments, it can also track personal identifiable information (PII).

The system provides necessary building blocks for telemetry. It includes an extensible event schema, APIs and libraries for Jupyter server, JupyterLab, JupyterHub and Jupyter Classic, and a flexible configuration system to work with any deployment.

The Jupyter Team remains committed to providing flexible and extensible building blocks for interactive computing. Tools like the Jupyter Notebook, JupyterLab, JupyterHub, etc. have expanded the world’s interactive computing toolkit. The Jupyter telemetry system offers new building blocks that improve these tools and increase security while protecting users’ privacy. Begin using these building blocks today. Telemetry is available [here](https://github.com/jupyter/telemetry).
361 changes: 361 additions & 0 deletions 30-telemetry/proposal.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,361 @@
# Jupyter telemetry enhancement proposal

| Item | Value |
|------------|------------------------------------------------------------------------------------------------------------------------------|
| JEP Number | 30 |
| Title | Jupyter telemetry |
| Authors | Jaipreet Singh ([email protected], @jaipreet-s), Zach Sailer([email protected], @Zsailer), Yuvi Panda([email protected], @yuvipanda) |
| Status | Draft |
| Type | S - [Standards Track](https://www.python.org/dev/peps/#pep-types-key) JEP |
| Created | 5 July 2019 |
| History | 5 July 2019

## Problem

Telemetry data serves a number of purposes for organizations deploying Jupyter: i) operational monitoring, ii) security monitoring / intrusion detection, iii) compliance auditing, and iv) a-posteriori analysis of the platform’s usage and design–i.e. as a UX research tool. This must be done in a transparent and understandable manner for end-users. Jupyter components and extension developers interested in publishing events must be incentivized to follow the system's interface and patterns rather then rolling out their own.

### Personas

* **Jupyter Users** are the primary stakeholder since it is their data.
* **Jupyter extension developers** will be emitting events from both the server-side and the browser-side code.
* **Jupyter operators** handle the infrastructure deployments and are interested in collecting various bits of events.
* **Analysts** consume the final stored data and visualize and interpret it.

### User Stories

1. As a Jupyter user, I can view and understand the events being collected, so that I have visibility into what information is being collected.
2. As a Jupyter user, I may be able to opt-in to event collection, so that I can choose what data about my usage is gathered.
3. As a Jupyter extension developer, I can publish custom events using a well defined interface, from the browser as well as the server.
4. As a Jupyter operator, I can configure events sinks for my deployment, so that I have control over where events are sent to.
5. As a Jupyter operator, I can write a custom event sink for my deployment, so that I can control where event data is stored.
6. As a Jupyter operator, I can decide what kinds and levels of events are going to be collected and stored.
7. As a Jupyter operator, I can configure opt-in or opt-out for the users in my deployment.
8. As an analyst, I can clearly tell what the fields in the data represent so that I can interpret and make decisions from that data.

## Proposed Enhancement

The enhancement spans two categories - general concepts applicable across all Jupyter components, and the implementation for each concept for each Jupyter component. This proposal covers the components required to publish and consume discrete events from the various components of the Jupyter ecosystem.

**General concepts:**

* Event schema: The schema for a given event which is respected by all the Jupyter components across the event's lifecycle.
* Event sinks: The "backends" which handle publishing the event to its final storage location.

**Jupyter components:**

* Jupyter server
* JupyterLab
* JupyterHub
* Jupyter Notebook (Classic)

## Detailed Explanation

### Event Schema

Each event is published conforming to a given [JSON schema](https://json-schema.org/) and this is the contract followed by each step in the event lifecycle. These are referred to by all publishers regardless of whether the event comes from the server or the browser.

Example event schema

```json
{
"name": "org.jupyter.kernel_lifecycle_event",
"version": 1,
"title": "Event emitted when a kernel lifecycle state changes",
"description": "Records each invocation of any command in JupyterLab",
"type": "object",
"args": {
"kernel_name": {
"type": "string",
"description": "Name of the kernel"
},
"state": {
"type": "string",
"description": "The new state of the kernel"
}
}
}
```

Schema validation is done in the core telemetry framework that routes events from publishers to sinks, so that each event sink does not need to implement schema validation.

Schemas are used at two places
* During the emit run-time, for validating events before publishing to the configured event sink.
* The schemas are picked up from the same package or a side-car package for runtime validation to avoid the network call.
* During analysis time, for discovering the available schemas and understanding the fields.
* Schema names are valid URIs. For public event discovery, these are resolvable, and private events could be just `com.mycompany.myevent1`

### Event Sinks

Event sinks are the backends where events published to. Event sinks can be configured from the browser as well as the server. Custom sinks can be implemented by extending the interface for the given Jupyter component.
See [Python interface](#python-event-sink-interface) and the [JupyterLab interface](#jupyterlab-event-sink-interface)

### Server-side components

#### Python publisher library

The Python event publisher library provides extension developers and other internal components an interface to record events. This is agnostic of the event sink back-end and the other deployment configurations.

```python
import jupyter_telemetry as telemetry

telemetry.record_event(
schema_name='org.jupyter.kernel_lifecycle_event',
schema_version='1',
args={
'kernel_name': 'python3',
'state': 'restarted'
}
)
```

#### Python Event Sink Interface

The event sink interface

```python
class MyCustomEventSink(JupyterEventSink):

def consume_event(event_data):
# Business logic to publish event into the backend.
```

#### Configuration

A setting on the Jupyter server allows operators to configure various options of the event sinks such as:

* The event sink implementations
* The list of event names to whitelist or blacklist

```json
c.NotebookApp.telemetry_event_sinks = [
"mypackage.sinks.MyCustomEventSink"
"mypackage.sinks.AnotherCustomEventSink"
]

c.NotebookApp.whitelisted_events = [
"org.jupyter.someevent"
]
```

#### Core event router

The implementation of the `telemetry.record_event` method which handles

* Schema validation
* Adds event metadata such as `creationTime`
* We should be able to add more metadata fields in the future if necessary without clashing with any potential schemas that have been defined in the various uses.
* Routing events to configured sinks
* Filters events based on the configuration
* (Optionally) Aggregation and data cleaning.

This part should be scalable to handle the volume of events and implement some sort of pub-sub design pattern.

```python
def record_event(name, schema_version, args):
# validate schema
# add metadata
# get configured sinks
# send event to sinks
```

#### REST Endpoint

In addition to the above sub-components, a REST interface is exposed on the Jupyter server to allow remote clients and the frontend to publish events into the server. The interface for this is similar to the [Python publisher library](#python-publisher-library)

```json
HTTP PUT /api/telemetry/event
{
"name": "org.jupyter.kernel_lifecycle_event",
"schema_version: "1",
"args": {
"kernel_name": "python3",
"state": "restarted"
}
}
```

#### Open Questions

1. This work is going to be done on the classic jupyter/notebook? What is the plan for porting to the standalone jupyter_server?

### JupyterHub

JupyterHub would use the same [underlying python library](#python-event-sink-interface)
as the notebook server, and be configured in the same way.

JupyterHub relies on extension through
[Spawners](https://jupyterhub.readthedocs.io/en/stable/reference/spawners.html),
[Authenticators](https://jupyterhub.readthedocs.io/en/stable/reference/authenticators.html),
[Proxies](https://jupyterhub.readthedocs.io/en/stable/api/spawner.html) &
[Services](https://jupyterhub.readthedocs.io/en/stable/reference/services.html)
to do most of its work. They would have interesting events to emit, so we should
make sure they have easy ways to.

#### Spawners, Authenticators & Proxies

Spawners, Authenticators & Proxies run in-process with JupyterHub, and have access to
the JupyterHub `app` object. When they want to emit events, they will:

1. Write schemas for events they want to emit
2. Register their schemas with the with the eventlogging library

```python
self.app.event_log.register_schema
for schema_file in glob(os.path.join(here, 'event-schemas','*.json')):
with open(schema_file) as f:
self.event_log.register_schema(json.load(f))
```
3. Emit events wherever they need.

```python
self.app.event_log.record_event('kubespawner.hub.jupyter.org/pod-missing', 1, {
'pod_name': pod.metadata.name,
'node': pod.metadata.node
})
```

This would get routed to the appropriate sink if the schema is whitelisted.

#### Services

Services run as their own process, and hence do not have access to the JupyterHub
`app` object. They should use the core eventlogging library directly, and admins
should be able to configure it as they would a standalone application.

#### Authenticated routing service
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This would be a "parallel universe" to https://github.com/jupyter/enhancement-proposals/pull/41/files#diff-5c74b6c64dfb44b841261c64623c9c6eR140 right? As in as an extension I could send events to either of these and they'd end up in the same sinks?

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I'm going to defer to @yuvipanda on the JupyterHub functionality 😃


Events sent from users' notebook servers or client side interfaces directly to a
sink are untrusted - they are from user controlled code & can be anything. It
would be useful to provide a JupyterHub service that can validate that the
users are who they say they are - even though the rest of the event data
should be considered untrusted.
jaipreet-s marked this conversation as resolved.
Show resolved Hide resolved

This would expose a [REST Endpoint](#rest-endpoint) that can receive data as a sink from
other parts of the ecosystem (Notebook, JupyterLab, classic notebook,
other JupyterHub services). It would then add a metadata username field
to each event, based on the JupyterHub authentication information sent
alongside the request. The event will then be sent to the appropriate
sink configured for this service.

This is also helpful in smaller installations where there's no other
centralized event collection mechanism (like fluentd, stackdriver, etc).
Events can be sent here, and it can route them to an accessible location
(like a database, or the filesystem)

### JupyterLab

There are quite a few analytics frameworks that send events directly from the browser, so the round trip to the server can be avoided in certain deployments. Additionally, JupyterLab also publishes "platform" events which are subscribed to and published to the event sinks.

All the sub-components defined in the [Jupyter server](#jupyter-server) section are applicable here as well.

#### JupyterLab publisher library

An interface to expose to JupyterLab extension developers to be able to publish custom events. This is similar to [this](https://github.com/jupyterlab/jupyterlab-telemetry/)

```typescript
import {recordEvent} from '@jupyterlab/telemetry'

recordEvent({
name: 'org.jupyter.jupyterlab_command',
schema_version: '1',
args: {
'commandName': 'newFile'
}
})
```

#### JupyterLab Event Sink Interface

An interface to expose to operators writing their own custom sink (as JupyterLab extensions themselves) and register themselves with the core routing components. The default event sink implementation is to publish to the [Server REST Endpoint](#rest-endpoint).

```typescript
import {EventSink, EventData} from '@jupyterlab/telemetry'

export class MyCustomEventSink implements EventSink {
handle_event(eventData: EventData): Promise<void> {
// Business logic to publish event to backend.
}
}
```

#### JupyterLab Configuration

The ability to configure the event sinks to publish to as well as events to blacklist or whitelist events. This can likely be accomplished via JupyterLab settings and the JupyterLab dependency management mechanism.

The JupyterLab telemetry extension provides `Token<EventSink>` that event sink implementations can depend on and register themselves with the core event router.

#### JupyterLab Core event router

The implementation of the `@jupyterlab/telemetry/recordEvent` method which handles routing events to the configured sinks and additionally:

* Schema validation
* Adds event metadata such as `creationTime`
* Filters events based on the configuration
* (Optionally) Aggregation and data cleaning.

#### User opt-in

Since JupyterLab is the user facing component, it also contains UX features to give more visibility and transparency to the Jupyter user. In general, every component should make it clear to the user what data is being collected and possible ways to turn it off.

* UI for opting in or opting out of telemetry data collection
* UI for showing the list of events that are currently being collected.


### Jupyter Notebook (Classic) Frontend

The proposal for Jupyter Classis is to have a convenience JS library that can be used to pubish events to the server [REST Endpoint](#rest-endpoint).
This ensures that we provide support for Jupyter Classic but can rely on the Jupyter Server to do much of the heavy-lifting by relying on the Core Event Router, Event Sinks, and configuration done at the server level.

```javascript

var telemetry = require('jupyter-telemetry-js')

telemetry.record_event(
name='org.jupyter.kernel_lifecycle_event',
schema_version='1',
args={
'kernel_name': 'python3',
'state': 'restarted'
}
)
```

### Data protection

(This section needs to be filled out)

* What guidance do we provide? How do we provide this?
* What building blocks do we provide?

## Pros and Cons

PROS

* Implementation shares common principles and components across the Jupyter components.
* Decoupling between event publishers and event consumers.
* Flexible configuration for Jupyter operators to configure event sinks and event names.

CONS

* Current proposal does not have guidance and building blocks for compliance programs such as GDPR

## Appendix

### Tenets

(These are taken from [Brian's initial issue](https://github.com/jupyterlab/team-compass/issues/4).)

There are certainly ethical and legal questions around telemetry systems. To address these, the following tenets of the telemetry system are proposed:

* **Data protection by design and default.** Jupyter and its telemetry system should come with builtin tools that enable safe and secure deployments with all types of data. See [Art. 25 of GDPR](https://gdpr-info.eu/art-25-gdpr/) for details about this tenet.
* **Make it easy to do the right things by default.** There are many ways to collect and use telemetry data that are illegal and/or irresponsible. Jupyter’s telemetry system should encode best practices and make it easy for operators to be responsible and comply with relevant laws.
* **Choice in balancing tradeoffs.** There are two types of data JupyterLab: 1) the actual datasets users are working with in notebooks, and 2) telemetry data about the Jupyter users. At times, protecting these two types of data at the same time will require tradeoffs. For example, if a deployment is doing research with sensitive HIPAA or FERPA data, the operators need to closely monitor every action taken by its researchers using JupyterLab to ensure the sensitive data is used appropriately. At the same time, in some jurisdictions (EU) Jupyter users may be protected by local laws (GDPR) about what telemetry data can be recorded, how it can be used, the terms of that usage.
* **Don't ignore the need for telemetry**. Organizations deploying Jupyter need to collect telemetry for a range of purposes. If we ignore this need, they will route around the project, with potential legal and ethical complications. By being proactive, we can establish best practices and guardrails.

### References

0. Initial telemetry [JEP proposal](https://github.com/jupyter/telemetry/blob/master/JEP.md)
1. Original write up by ellisonbg@ [https://github.com/jupyterlab/team-compass/issues/4](https://github.com/jupyterlab/team-compass/issues/4)
2. Wikimedia [telemetry system](https://m.mediawiki.org/wiki/Extension:EventLogging/Guide)
3. Initial strawman [design doc](https://github.com/jupyterlab/jupyterlab-telemetry/blob/master/design.md)
4. [Mozilla Telemetry System](https://wiki.mozilla.org/Telemetry)