Net.cpp 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470
  1. /* Net.cpp
  2. *
  3. * Copyright (C) 2017,2018 Paul Boersma
  4. *
  5. * This code is free software; you can redistribute it and/or modify
  6. * it under the terms of the GNU General Public License as published by
  7. * the Free Software Foundation; either version 2 of the License, or (at
  8. * your option) any later version.
  9. *
  10. * This code is distributed in the hope that it will be useful, but
  11. * WITHOUT ANY WARRANTY; without even the implied warranty of
  12. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
  13. * See the GNU General Public License for more details.
  14. *
  15. * You should have received a copy of the GNU General Public License
  16. * along with this work. If not, see <http://www.gnu.org/licenses/>.
  17. */
  18. //#include <OpenCL/OpenCL.h>
  19. #include "Net.h"
  20. #include "oo_DESTROY.h"
  21. #include "Net_def.h"
  22. #include "oo_COPY.h"
  23. #include "Net_def.h"
  24. #include "oo_EQUAL.h"
  25. #include "Net_def.h"
  26. #include "oo_CAN_WRITE_AS_ENCODING.h"
  27. #include "Net_def.h"
  28. #include "oo_WRITE_TEXT.h"
  29. #include "Net_def.h"
  30. #include "oo_READ_TEXT.h"
  31. #include "Net_def.h"
  32. #include "oo_WRITE_BINARY.h"
  33. #include "Net_def.h"
  34. #include "oo_READ_BINARY.h"
  35. #include "Net_def.h"
  36. #include "oo_DESCRIPTION.h"
  37. #include "Net_def.h"
  38. Thing_implement (RBMLayer, Layer, 0);
  39. Thing_implement (FullyConnectedLayer, Layer, 0);
  40. Thing_implement (Net, Daata, 0);
  41. static autoRBMLayer RBMLayer_create (integer numberOfInputNodes, integer numberOfOutputNodes, bool inputsAreBinary) {
  42. try {
  43. autoRBMLayer me = Thing_new (RBMLayer);
  44. my numberOfInputNodes = numberOfInputNodes;
  45. my inputBiases = VECzero (numberOfInputNodes);
  46. my inputActivities = VECzero (numberOfInputNodes);
  47. my inputReconstruction = VECzero (numberOfInputNodes);
  48. my numberOfOutputNodes = numberOfOutputNodes;
  49. my outputBiases = VECzero (numberOfOutputNodes);
  50. my outputActivities = VECzero (numberOfOutputNodes);
  51. my outputReconstruction = VECzero (numberOfOutputNodes);
  52. my weights = MATzero (numberOfInputNodes, numberOfOutputNodes);
  53. my inputsAreBinary = inputsAreBinary;
  54. return me;
  55. } catch (MelderError) {
  56. Melder_throw (U"RBM layer with ", numberOfInputNodes, U" input nodes and ",
  57. numberOfOutputNodes, U" output nodes not created.");
  58. }
  59. }
  60. autoNet Net_createEmpty (integer numberOfInputNodes) {
  61. try {
  62. autoNet me = Thing_new (Net);
  63. //my numberOfInputNodes = numberOfInputNodes;
  64. return me;
  65. } catch (MelderError) {
  66. Melder_throw (U"Net not created.");
  67. }
  68. }
  69. void Net_initAsDeepBeliefNet (Net me, constVEC numbersOfNodes, bool inputsAreBinary) {
  70. if (numbersOfNodes.size < 2)
  71. Melder_throw (U"A deep belief net should have at least two levels of nodes.");
  72. integer numberOfLayers = numbersOfNodes.size - 1;
  73. my layers = LayerList_create ();
  74. for (integer ilayer = 1; ilayer <= numberOfLayers; ilayer ++) {
  75. autoRBMLayer layer = RBMLayer_create (
  76. Melder_iround (numbersOfNodes [ilayer]),
  77. Melder_iround (numbersOfNodes [ilayer + 1]),
  78. ilayer == 1 ? inputsAreBinary : true
  79. );
  80. my layers -> addItem_move (layer.move());
  81. }
  82. }
  83. autoNet Net_createAsDeepBeliefNet (constVEC numbersOfNodes, bool inputsAreBinary) {
  84. try {
  85. autoNet me = Thing_new (Net);
  86. Net_initAsDeepBeliefNet (me.get(), numbersOfNodes, inputsAreBinary);
  87. return me;
  88. } catch (MelderError) {
  89. Melder_throw (U"Net not created.");
  90. }
  91. }
  92. static void copyOutputsToInputs (Layer me, Layer you) {
  93. vectorcopy_preallocated (your inputActivities.get(), my outputActivities.get());
  94. }
  95. inline static double logistic (double excitation) {
  96. return 1.0 / (1.0 + exp (- excitation));
  97. }
  98. static void Layer_sampleOutput (Layer me) {
  99. for (integer jnode = 1; jnode <= my numberOfOutputNodes; jnode ++) {
  100. double probability = my outputActivities [jnode];
  101. my outputActivities [jnode] = (double) NUMrandomBernoulli (probability);
  102. }
  103. }
  104. void structRBMLayer :: v_spreadUp (kLayer_activationType activationType) {
  105. integer numberOfOutputNodes = our numberOfOutputNodes;
  106. for (integer jnode = 1; jnode <= numberOfOutputNodes; jnode ++) {
  107. PAIRWISE_SUM (longdouble, excitation, integer, our numberOfInputNodes,
  108. double *p_inputActivity = & our inputActivities [1];
  109. double *p_weight = & our weights [1] [jnode],
  110. (longdouble) *p_inputActivity * (longdouble) *p_weight,
  111. ( p_inputActivity += 1, p_weight += numberOfOutputNodes )
  112. )
  113. excitation += our outputBiases [jnode];
  114. our outputActivities [jnode] = logistic ((double) excitation);
  115. }
  116. if (activationType == kLayer_activationType::STOCHASTIC)
  117. Layer_sampleOutput (this);
  118. }
  119. void Net_spreadUp (Net me, kLayer_activationType activationType) {
  120. for (integer ilayer = 1; ilayer <= my layers->size; ilayer ++) {
  121. Layer layer = my layers->at [ilayer];
  122. if (ilayer > 1)
  123. copyOutputsToInputs (my layers->at [ilayer - 1], layer);
  124. layer -> v_spreadUp (activationType);
  125. }
  126. }
  127. void structRBMLayer :: v_sampleInput () {
  128. for (integer inode = 1; inode <= our numberOfInputNodes; inode ++) {
  129. if (our inputsAreBinary) {
  130. double probability = our inputActivities [inode];
  131. our inputActivities [inode] = (double) NUMrandomBernoulli (probability);
  132. } else { // Gaussian
  133. double excitation = our inputActivities [inode];
  134. our inputActivities [inode] = NUMrandomGauss (excitation, 1.0);
  135. }
  136. }
  137. }
  138. void Net_sampleInput (Net me) {
  139. my layers->at [1] -> v_sampleInput ();
  140. }
  141. void Net_sampleOutput (Net me) {
  142. Layer_sampleOutput (my layers->at [my layers->size]);
  143. }
  144. static void copyInputsToOutputs (Layer me, Layer you) {
  145. vectorcopy_preallocated (your outputActivities.get(), my inputActivities.get());
  146. }
  147. void structRBMLayer :: v_spreadDown (kLayer_activationType activationType) {
  148. for (integer inode = 1; inode <= our numberOfInputNodes; inode ++) {
  149. PAIRWISE_SUM (longdouble, excitation, integer, our numberOfOutputNodes,
  150. double *p_weight = & our weights [inode] [1];
  151. double *p_outputActivity = & our outputActivities [1],
  152. (longdouble) *p_weight * (longdouble) *p_outputActivity,
  153. ( p_weight += 1, p_outputActivity += 1 )
  154. )
  155. excitation += our inputBiases [inode];
  156. if (our inputsAreBinary) {
  157. our inputActivities [inode] = logistic ((double) excitation);
  158. } else { // linear
  159. our inputActivities [inode] = (double) excitation;
  160. }
  161. }
  162. if (activationType == kLayer_activationType::STOCHASTIC)
  163. our v_sampleInput ();
  164. }
  165. void Net_spreadDown (Net me, kLayer_activationType activationType) {
  166. for (integer ilayer = my layers->size; ilayer > 0; ilayer --) {
  167. Layer layer = my layers->at [ilayer];
  168. if (ilayer < my layers->size)
  169. copyInputsToOutputs (my layers->at [ilayer + 1], layer);
  170. layer -> v_spreadDown (activationType);
  171. }
  172. }
  173. void structRBMLayer :: v_spreadDown_reconstruction () {
  174. for (integer inode = 1; inode <= our numberOfInputNodes; inode ++) {
  175. PAIRWISE_SUM (longdouble, excitation, integer, our numberOfOutputNodes,
  176. double *p_weight = & our weights [inode] [1];
  177. double *p_outputActivity = & our outputActivities [1],
  178. (longdouble) *p_weight * (longdouble) *p_outputActivity,
  179. ( p_weight += 1, p_outputActivity += 1 )
  180. )
  181. excitation += our inputBiases [inode];
  182. if (our inputsAreBinary) {
  183. our inputReconstruction [inode] = logistic ((double) excitation);
  184. } else { // linear
  185. our inputReconstruction [inode] = (double) excitation;
  186. }
  187. }
  188. }
  189. void Net_spreadDown_reconstruction (Net me) {
  190. for (integer ilayer = my layers->size; ilayer > 0; ilayer --) {
  191. my layers->at [ilayer] -> v_spreadDown_reconstruction ();
  192. }
  193. }
  194. void structRBMLayer :: v_spreadUp_reconstruction () {
  195. integer numberOfOutputNodes = our numberOfOutputNodes;
  196. for (integer jnode = 1; jnode <= our numberOfOutputNodes; jnode ++) {
  197. PAIRWISE_SUM (longdouble, excitation, integer, our numberOfInputNodes,
  198. double *p_inputActivity = & our inputReconstruction [1];
  199. double *p_weight = & our weights [1] [jnode],
  200. (longdouble) *p_inputActivity * (longdouble) *p_weight,
  201. ( p_inputActivity += 1, p_weight += numberOfOutputNodes )
  202. )
  203. excitation += our outputBiases [jnode];
  204. our outputReconstruction [jnode] = logistic ((double) excitation);
  205. }
  206. }
  207. void Net_spreadUp_reconstruction (Net me) {
  208. for (integer ilayer = 1; ilayer <= my layers->size; ilayer ++) {
  209. my layers->at [ilayer] -> v_spreadUp_reconstruction ();
  210. }
  211. }
  212. void structRBMLayer :: v_update (double learningRate) {
  213. for (integer jnode = 1; jnode <= our numberOfOutputNodes; jnode ++) {
  214. our outputBiases [jnode] += learningRate * (our outputActivities [jnode] - our outputReconstruction [jnode]);
  215. }
  216. for (integer inode = 1; inode <= our numberOfInputNodes; inode ++) {
  217. our inputBiases [inode] += learningRate * (our inputActivities [inode] - our inputReconstruction [inode]);
  218. for (integer jnode = 1; jnode <= our numberOfOutputNodes; jnode ++) {
  219. our weights [inode] [jnode] += learningRate *
  220. (our inputActivities [inode] * our outputActivities [jnode] -
  221. our inputReconstruction [inode] * our outputReconstruction [jnode]);
  222. }
  223. }
  224. }
  225. void Net_update (Net me, double learningRate) {
  226. for (integer ilayer = 1; ilayer <= my layers->size; ilayer ++) {
  227. my layers->at [ilayer] -> v_update (learningRate);
  228. }
  229. }
  230. static void Layer_PatternList_applyToInput (Layer me, PatternList thee, integer rowNumber) {
  231. Melder_require (my numberOfInputNodes == thy nx,
  232. U"The number of elements in each row of ", thee, U" (", thy nx,
  233. U") does not match the number of input nodes of ", me, U" (", my numberOfInputNodes, U").");
  234. for (integer ifeature = 1; ifeature <= my numberOfInputNodes; ifeature ++) {
  235. my inputActivities [ifeature] = thy z [rowNumber] [ifeature];
  236. }
  237. }
  238. void Net_PatternList_applyToInput (Net me, PatternList thee, integer rowNumber) {
  239. try {
  240. Layer_PatternList_applyToInput (my layers->at [1], thee, rowNumber);
  241. } catch (MelderError) {
  242. Melder_throw (me, U" & ", thee, U": pattern ", rowNumber, U" not applied to input.");
  243. }
  244. }
  245. static void Layer_PatternList_applyToOutput (Layer me, PatternList thee, integer rowNumber) {
  246. Melder_require (my numberOfOutputNodes == thy nx,
  247. U"The number of elements in each row of ", thee, U" (", thy nx,
  248. U") does not match the number of output nodes of ", me, U" (", my numberOfOutputNodes, U").");
  249. for (integer icat = 1; icat <= my numberOfOutputNodes; icat ++) {
  250. my outputActivities [icat] = thy z [rowNumber] [icat];
  251. }
  252. }
  253. void Net_PatternList_applyToOutput (Net me, PatternList thee, integer rowNumber) {
  254. Layer_PatternList_applyToOutput (my layers->at [my layers->size], thee, rowNumber);
  255. }
  256. void Net_PatternList_learn (Net me, PatternList thee, double learningRate) {
  257. try {
  258. for (integer ipattern = 1; ipattern <= thy ny; ipattern ++) {
  259. Net_PatternList_applyToInput (me, thee, ipattern);
  260. Net_spreadUp (me, kLayer_activationType::STOCHASTIC);
  261. for (integer ilayer = 1; ilayer <= my layers->size; ilayer ++) {
  262. Layer layer = my layers->at [ilayer];
  263. layer -> v_spreadDown_reconstruction ();
  264. layer -> v_spreadUp_reconstruction ();
  265. layer -> v_update (learningRate);
  266. }
  267. }
  268. } catch (MelderError) {
  269. Melder_throw (me, U" & ", thee, U": not learned.");
  270. }
  271. }
  272. void Net_PatternList_learnByLayer (Net me, PatternList thee, double learningRate) {
  273. try {
  274. for (integer ilayer = 1; ilayer <= my layers->size; ilayer ++) {
  275. Layer layer = my layers->at [ilayer];
  276. for (integer ipattern = 1; ipattern <= thy ny; ipattern ++) {
  277. Layer_PatternList_applyToInput (my layers->at [1], thee, ipattern);
  278. my layers->at [1] -> v_spreadUp (kLayer_activationType::STOCHASTIC);
  279. for (integer jlayer = 2; jlayer <= ilayer; jlayer ++) {
  280. copyOutputsToInputs (my layers->at [jlayer - 1], my layers->at [jlayer]);
  281. my layers->at [jlayer] -> v_spreadUp (kLayer_activationType::STOCHASTIC);
  282. }
  283. layer -> v_spreadDown_reconstruction ();
  284. layer -> v_spreadUp_reconstruction ();
  285. layer -> v_update (learningRate);
  286. }
  287. }
  288. } catch (MelderError) {
  289. Melder_throw (me, U" & ", thee, U": not learned.");
  290. }
  291. }
  292. autoActivationList Net_PatternList_to_ActivationList (Net me, PatternList thee, kLayer_activationType activationType) {
  293. try {
  294. Layer outputLayer = my layers->at [my layers->size];
  295. autoActivationList activations = ActivationList_create (thy ny, outputLayer -> numberOfOutputNodes);
  296. for (integer ipattern = 1; ipattern <= thy ny; ipattern ++) {
  297. Net_PatternList_applyToInput (me, thee, ipattern);
  298. Net_spreadUp (me, activationType);
  299. NUMvector_copyElements <double> (outputLayer -> outputActivities.at, activations -> z [ipattern], 1, outputLayer -> numberOfOutputNodes);
  300. }
  301. return activations;
  302. } catch (MelderError) {
  303. Melder_throw (me, thee, U"No ActivationList created.");
  304. }
  305. }
  306. static autoMatrix Layer_extractInputActivities (Layer me) {
  307. try {
  308. autoMatrix thee = Matrix_createSimple (1, my numberOfInputNodes);
  309. NUMvector_copyElements <double> (my inputActivities.at, thy z [1], 1, my numberOfInputNodes);
  310. return thee;
  311. } catch (MelderError) {
  312. Melder_throw (me, U": input activities not extracted.");
  313. }
  314. }
  315. autoMatrix Net_extractInputActivities (Net me) {
  316. return Layer_extractInputActivities (my layers->at [1]);
  317. }
  318. static autoMatrix Layer_extractOutputActivities (Layer me) {
  319. try {
  320. autoMatrix thee = Matrix_createSimple (1, my numberOfOutputNodes);
  321. NUMvector_copyElements <double> (my outputActivities.at, thy z [1], 1, my numberOfOutputNodes);
  322. return thee;
  323. } catch (MelderError) {
  324. Melder_throw (me, U": output activities not extracted.");
  325. }
  326. }
  327. autoMatrix Net_extractOutputActivities (Net me) {
  328. return Layer_extractOutputActivities (my layers->at [my layers->size]);
  329. }
  330. autoMatrix structRBMLayer :: v_extractInputReconstruction () {
  331. try {
  332. autoMatrix thee = Matrix_createSimple (1, our numberOfInputNodes);
  333. NUMvector_copyElements <double> (our inputReconstruction.at, thy z [1], 1, our numberOfInputNodes);
  334. return thee;
  335. } catch (MelderError) {
  336. Melder_throw (this, U": input reconstruction not extracted.");
  337. }
  338. }
  339. autoMatrix Net_extractInputReconstruction (Net me) {
  340. return my layers->at [1] -> v_extractInputReconstruction ();
  341. }
  342. autoMatrix structRBMLayer :: v_extractOutputReconstruction () {
  343. try {
  344. autoMatrix thee = Matrix_createSimple (1, our numberOfOutputNodes);
  345. NUMvector_copyElements <double> (our outputReconstruction.at, thy z [1], 1, our numberOfOutputNodes);
  346. return thee;
  347. } catch (MelderError) {
  348. Melder_throw (this, U": output reconstruction not extracted.");
  349. }
  350. }
  351. autoMatrix Net_extractOutputReconstruction (Net me) {
  352. return my layers->at [my layers->size] -> v_extractOutputReconstruction ();
  353. }
  354. autoMatrix structRBMLayer :: v_extractInputBiases () {
  355. try {
  356. autoMatrix thee = Matrix_createSimple (1, our numberOfInputNodes);
  357. NUMvector_copyElements <double> (our inputBiases.at, thy z [1], 1, our numberOfInputNodes);
  358. return thee;
  359. } catch (MelderError) {
  360. Melder_throw (this, U": input biases not extracted.");
  361. }
  362. }
  363. static void Net_checkLayerNumber (Net me, integer layerNumber) {
  364. Melder_require (layerNumber >= 1,
  365. U"Your layer number (", layerNumber, U") should be positive.");
  366. Melder_require (layerNumber <= my layers->size,
  367. U"Your layer number (", layerNumber, U") should be at most my number of layers (", my layers->size, U").");
  368. }
  369. autoMatrix Net_extractInputBiases (Net me, integer layerNumber) {
  370. try {
  371. Net_checkLayerNumber (me, layerNumber);
  372. return my layers->at [layerNumber] -> v_extractInputBiases ();
  373. } catch (MelderError) {
  374. Melder_throw (me, U": input biases not extracted from layer (", layerNumber, U").");
  375. }
  376. }
  377. autoMatrix structRBMLayer :: v_extractOutputBiases () {
  378. try {
  379. autoMatrix thee = Matrix_createSimple (1, our numberOfOutputNodes);
  380. NUMvector_copyElements <double> (our outputBiases.at, thy z [1], 1, our numberOfOutputNodes);
  381. return thee;
  382. } catch (MelderError) {
  383. Melder_throw (this, U": input biases not extracted.");
  384. }
  385. }
  386. autoMatrix Net_extractOutputBiases (Net me, integer layerNumber) {
  387. try {
  388. Net_checkLayerNumber (me, layerNumber);
  389. return my layers->at [layerNumber] -> v_extractOutputBiases ();
  390. } catch (MelderError) {
  391. Melder_throw (me, U": output biases not extracted from layer (", layerNumber, U").");
  392. }
  393. }
  394. autoMatrix structRBMLayer :: v_extractWeights () {
  395. try {
  396. autoMatrix thee = Matrix_createSimple (our numberOfInputNodes, our numberOfOutputNodes);
  397. matrixcopy_preallocated (thy z.get(), our weights.get());
  398. return thee;
  399. } catch (MelderError) {
  400. Melder_throw (this, U": weights not extracted.");
  401. }
  402. }
  403. autoMatrix Net_extractWeights (Net me, integer layerNumber) {
  404. try {
  405. Net_checkLayerNumber (me, layerNumber);
  406. return my layers->at [layerNumber] -> v_extractWeights ();
  407. } catch (MelderError) {
  408. Melder_throw (me, U": weights not extracted from layer (", layerNumber, U").");
  409. }
  410. }
  411. autoMAT structRBMLayer :: v_getWeights () {
  412. return matrixcopy (our weights.get());
  413. }
  414. autoMAT Net_getWeights (Net me, integer layerNumber) {
  415. return my layers->at [layerNumber] -> v_getWeights ();
  416. }
  417. /* End of file Net.cpp */