"# Fitting a model to data with both x and y errors with `Bilby`\n",
"\n",
"Usually when we fit a model to data with a Gaussian Likelihood we assume that we know x values exactly. This is almost never the case. Here we show how to fit a model with errors in both x and y.\n",
"\n",
"Since we are using a very simple model we will use the `nestle` sampler.\n",
"This can be installed using\n",
"\n",
"```console\n",
"$ conda install -c conda-forge nestle\n",
"```\n",
"\n",
"or\n",
"\n",
"```console\n",
"$ pip install nestle\n",
"```"
"Usually when we fit a model to data with a Gaussian Likelihood we assume that we know x values exactly. This is almost never the case. Here we show how to fit a model with errors in both x and y."
# Fitting a model to data with both x and y errors with `Bilby`
Usually when we fit a model to data with a Gaussian Likelihood we assume that we know x values exactly. This is almost never the case. Here we show how to fit a model with errors in both x and y.
Since we are using a very simple model we will use the `nestle` sampler.
Our first step is to recover the straight line using a simple Gaussian Likelihood that only takes into account the y errors. Under the assumption we know x exactly. In this case, we pass in xtrue for x
### Fit with unmodeled uncertainty in the x-values
As expected this is easy to recover and the sampler does a good job. However this was made too easy - by passing in the 'true' values of x. Lets see what happens when we pass in the observed values of x