IEEE Trans Cybern. 2014 Nov;44(11):2202-20. doi: 10.1109/TCYB.2014.2303712.
Stochastic point location (SPL) deals with the problem of a learning mechanism (LM) determining the optimal point on the line when the only input it receives are stochastic signals about the direction in which it should move. One can differentiate the SPL from the traditional class of optimization problems by the fact that the former considers the case where the directional information, for example, as inferred from an Oracle (which possibly computes the derivatives), suffices to achieve the optimization-without actually explicitly computing any derivatives. The SPL can be described in terms of a LM (algorithm) attempting to locate a point on a line. The LM interacts with a random environment which essentially informs it, possibly erroneously, if the unknown parameter is on the left or the right of a given point. Given a current estimate of the optimal solution, all the reported solutions to this problem effectively move along the line to yield updated estimates which are in the neighborhood of the current solution(1) This paper proposes a dramatically distinct strategy, namely, that of partitioning the line in a hierarchical tree-like manner, and of moving to relatively distant points, as characterized by those along the path of the tree. We are thus attempting to merge the rich fields of stochastic optimization and data structures. Indeed, as in the original discretized solution to the SPL, in one sense, our solution utilizes the concept of discretization and operates a uni-dimensional controlled random walk (RW) in the discretized space, to locate the unknown parameter. However, by moving to nonneighbor points in the space, our newly proposed hierarchical stochastic searching on the line (HSSL) solution performs such a controlled RW on the discretized space structured on a superimposed binary tree. We demonstrate that the HSSL solution is orders of magnitude faster than the original SPL solution proposed by Oommen. By a rigorous analysis, the HSSL is shown to be optimal if the effectiveness (or credibility) of the environment, given by p , is greater than the golden ratio conjugate. The solution has been both analytically solved and simulated, and the results obtained are extremely fascinating, as this is the first reported use of time reversibility in the analysis of stochastic learning. The learning automata extensions of the scheme are currently being investigated. As we shall see later, hierarchical solutions have been proposed in the field of LA.
随机点定位(SPL)处理的是学习机制(LM)在接收到有关其移动方向的随机信号时,确定线上最佳点的问题。可以通过以下事实将 SPL 与传统优化问题区分开来:前者考虑了这样一种情况,即定向信息(例如,从可能计算导数的 Oracle 推断得出)足以实现优化,而无需实际明确计算任何导数。SPL 可以用试图在直线上定位点的 LM(算法)来描述。LM 与随机环境交互,该环境本质上会错误地通知它未知参数是在给定点的左侧还是右侧。给定最佳解的当前估计,该问题的所有报告解决方案都有效地在线上移动,从而产生接近当前解决方案的更新估计(1)本文提出了一种截然不同的策略,即层次树状方式分割线,并移动到相对较远的点,其特征是沿树路径的那些点。因此,我们试图融合随机优化和数据结构这两个领域。事实上,就 SPL 的原始离散解决方案而言,从某种意义上说,我们的解决方案利用了离散化的概念,并在离散化空间中进行一维受控随机游走(RW),以定位未知参数。然而,通过移动到空间中的非邻域点,我们新提出的在线分层随机搜索(HSSL)解决方案在叠加二叉树上的离散化空间中执行这种受控 RW。我们证明 HSSL 解决方案比 Oommen 提出的原始 SPL 解决方案快几个数量级。通过严格的分析,当环境的有效性(或可信度)由 p 给出时,HSSL 是最优的,如果大于黄金比例共轭。该解决方案已经进行了分析和模拟,结果非常吸引人,因为这是首次在随机学习分析中报告使用时间可逆性。该方案的学习自动机扩展目前正在研究中。正如我们稍后将看到的,分层解决方案已经在 LA 领域提出。