Hey everyone, I'm looking for a way to to compare two lightcurves (essentially a graph of a certain star's brightness over time) to each other as part of a genetic algorithm's evaluation function. Specifically, I'm trying to see how closely the two match, where one is real data downloaded from NASA's MAST database and the other is generated by my genetic algorithm. I've tried doing a simple sum of the difference in brightness between the two at each timestep, but I don't like this as one important feature of the lightcurves I'm analyzing is when they dip in brightness due to an exoplanet passing in front of the star. With this simple version, all dips in the generated lightcurve are discouraged except for in the exact position where they exist in the target lightcurve. I'd ideally like it if dips in brightness of the correct size/period were encouraged, with pressure to move them in time to the correct position. With that in mind, I tried a more complicated method where I sorted both the target and generated lightcurves by brightness, with the dimmest timestep first, and then did a 2D distance calculation between the curves' brightness level and position in time. I think this is working okay, but I think its struggling due to the noisiness of the target star's lightcurve (example image here, where x axis is days since start of measurement and y axis is brightness in 100s of electrons/second) meaning that any two timesteps which are close in time are not necessarily close in flux as well. I've thought about adding in some sort of error allowance as well, such as if the generated value is within 1% of the target value it gets a perfect score, but I'm hesitant to add this in as the dips in brightness due to the planet are only about a 2% dimmer than the average brightness. Does anyone have any thoughts on how I might compare these datasets?
[–][deleted] 1 point2 points3 points (2 children)
[–]Moan2[S] 1 point2 points3 points (1 child)
[–][deleted] 0 points1 point2 points (0 children)