You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Second, using the natural time when samples arrive instead of a surrogate time is required sometimes. DBSTREAM does not distinguish 2 scenarios when samples come at
100ms and 200ms
100ms and 500ms
if these arrivals are sequential (no other samples arrive in between). However, there can be a large difference from the business perspective.
I propose to add a t (time) argument to the learn_one function. Then, we can learn from samples using the same t value if samples arrive simultaneously and supply time in any units to this function, adjusting the fading_factor.
The text was updated successfully, but these errors were encountered:
@ShkarupaDC Hi! Sorry for getting this late to get back to you.
Within the original paper, the authors have designed DBSTREAM with the time step concept. This means that from my understanding, the authors only consider the order of which data comes, not the speed at which data comes. As such, this is the reason why I use time_stamp in my implementation to represent the parameter t in the original paper.
Moreover, in data stream in general, we usually assume that data comes once at a time. As such, when samples arrive simultaneously, what we would usually do would be to consider them as other data points, coming one by one and in order. This might seem unreasonable, but to make DBSTREAM compatible with the design language of River in general, and to align with such philosophy, we decided to implement it this way.
Hope that this answer clearly explains your concerns.
Hi! I want to propose a new feature for DBSTREAM.
DBSTREAM uses a protected internal timer (
_time_stamp
) to measure the time between learning steps. There are 2 issues with this approach.if these arrivals are sequential (no other samples arrive in between). However, there can be a large difference from the business perspective.
I propose to add a
t
(time) argument to thelearn_one
function. Then, we can learn from samples using the samet
value if samples arrive simultaneously and supply time in any units to this function, adjusting thefading_factor
.The text was updated successfully, but these errors were encountered: