tx · AwRrLAxF5aEBEdiTUheEyimrHT9gq5mRd1ki8zFuTWZw

3N3n75UqB8G1GKmXFr4zPhKCjGcqJPRSuJY:  -0.01000000 Waves

2024.04.28 11:30 [3082469] smart account 3N3n75UqB8G1GKmXFr4zPhKCjGcqJPRSuJY > SELF 0.00000000 Waves

{ "type": 13, "id": "AwRrLAxF5aEBEdiTUheEyimrHT9gq5mRd1ki8zFuTWZw", "fee": 1000000, "feeAssetId": null, "timestamp": 1714293015946, "version": 2, "chainId": 84, "sender": "3N3n75UqB8G1GKmXFr4zPhKCjGcqJPRSuJY", "senderPublicKey": "2AWdnJuBMzufXSjTvzVcawBQQhnhF1iXR6QNVgwn33oc", "proofs": [ "TT3MU9AZKuxbWrB4YBVdaN9ZdEprtRfc2HrZDWesq5XpEkaoh1H46dQhNuT3pDhT2qsFWcsnLwRdng9DMb6mDBj" ], "script": "base64:AAIFAAAAAAAAAAgIAhIECgIBAQAAAAgAAAAADWxheWVyMVdlaWdodHMJAARMAAAAAgkABEwAAAACAAAAAAAACSmxCQAETAAAAAIAAAAAAAAJKpwFAAAAA25pbAkABEwAAAACCQAETAAAAAIAAAAAAAAGUfUJAARMAAAAAgAAAAAAAAZSLQUAAAADbmlsBQAAAANuaWwAAAAADGxheWVyMUJpYXNlcwkABEwAAAACAP///////AwVCQAETAAAAAIA///////2TQsFAAAAA25pbAAAAAANbGF5ZXIyV2VpZ2h0cwkABEwAAAACCQAETAAAAAIAAAAAAAAMtcUJAARMAAAAAgD///////JPigUAAAADbmlsBQAAAANuaWwAAAAADGxheWVyMkJpYXNlcwkABEwAAAACAP//////+i8FBQAAAANuaWwBAAAAEWV4cF9hcHByb3hpbWF0aW9uAAAAAQAAAAF4BAAAAAVzY2FsZQAAAAAAAAAnEAQAAAAIc2NhbGVkX3gJAABpAAAAAgUAAAABeAUAAAAFc2NhbGUEAAAAAWUAAAAAAAAAai8EAAAAB2ZhY3RvcjEFAAAACHNjYWxlZF94BAAAAAdmYWN0b3IyCQAAaQAAAAIJAABoAAAAAgUAAAAIc2NhbGVkX3gFAAAACHNjYWxlZF94CQAAaAAAAAIAAAAAAAAAAAIFAAAABXNjYWxlBAAAAAdmYWN0b3IzCQAAaQAAAAIJAABoAAAAAgkAAGgAAAACBQAAAAhzY2FsZWRfeAUAAAAIc2NhbGVkX3gFAAAACHNjYWxlZF94CQAAaAAAAAIJAABoAAAAAgAAAAAAAAAABgUAAAAFc2NhbGUFAAAABXNjYWxlCQAAZAAAAAIJAABkAAAAAgkAAGQAAAACAAAAAAAAACcQBQAAAAdmYWN0b3IxBQAAAAdmYWN0b3IyBQAAAAdmYWN0b3IzAQAAAAdzaWdtb2lkAAAAAgAAAAF6AAAAC2RlYnVnUHJlZml4BAAAAARiYXNlAAAAAAAAACcQBAAAAAlwb3NpdGl2ZVoDCQAAZgAAAAIAAAAAAAAAAAAFAAAAAXoJAQAAAAEtAAAAAQUAAAABegUAAAABegQAAAAIZXhwVmFsdWUJAQAAABFleHBfYXBwcm94aW1hdGlvbgAAAAEFAAAACXBvc2l0aXZlWgQAAAAIc2lnVmFsdWUJAABpAAAAAgkAAGgAAAACBQAAAARiYXNlAAAAAAAAACcQCQAAZAAAAAIFAAAABGJhc2UFAAAACGV4cFZhbHVlCQAFFAAAAAIJAARMAAAAAgkBAAAADEludGVnZXJFbnRyeQAAAAIJAAEsAAAAAgUAAAALZGVidWdQcmVmaXgCAAAACXBvc2l0aXZlWgUAAAAJcG9zaXRpdmVaCQAETAAAAAIJAQAAAAxJbnRlZ2VyRW50cnkAAAACCQABLAAAAAIFAAAAC2RlYnVnUHJlZml4AgAAAAhleHBWYWx1ZQUAAAAIZXhwVmFsdWUJAARMAAAAAgkBAAAADEludGVnZXJFbnRyeQAAAAIJAAEsAAAAAgUAAAALZGVidWdQcmVmaXgCAAAACHNpZ1ZhbHVlBQAAAAhzaWdWYWx1ZQUAAAADbmlsBQAAAAhzaWdWYWx1ZQEAAAARZm9yd2FyZFBhc3NMYXllcjEAAAAEAAAABWlucHV0AAAAB3dlaWdodHMAAAAGYmlhc2VzAAAAC2RlYnVnUHJlZml4BAAAAARzdW0wCQAAZAAAAAIJAABkAAAAAgkAAGsAAAADCQABkQAAAAIFAAAABWlucHV0AAAAAAAAAAAACQABkQAAAAIJAAGRAAAAAgUAAAAHd2VpZ2h0cwAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAA9CQAkAAGsAAAADCQABkQAAAAIFAAAABWlucHV0AAAAAAAAAAABCQABkQAAAAIJAAGRAAAAAgUAAAAHd2VpZ2h0cwAAAAAAAAAAAAAAAAAAAAAAAQAAAAAAAA9CQAkAAZEAAAACBQAAAAZiaWFzZXMAAAAAAAAAAAAEAAAABHN1bTEJAABkAAAAAgkAAGQAAAACCQAAawAAAAMJAAGRAAAAAgUAAAAFaW5wdXQAAAAAAAAAAAAJAAGRAAAAAgkAAZEAAAACBQAAAAd3ZWlnaHRzAAAAAAAAAAABAAAAAAAAAAAAAAAAAAAAD0JACQAAawAAAAMJAAGRAAAAAgUAAAAFaW5wdXQAAAAAAAAAAAEJAAGRAAAAAgkAAZEAAAACBQAAAAd3ZWlnaHRzAAAAAAAAAAABAAAAAAAAAAABAAAAAAAAD0JACQABkQAAAAIFAAAABmJpYXNlcwAAAAAAAAAAAQQAAAALJHQwMTY3MjE3MTgJAQAAAAdzaWdtb2lkAAAAAgUAAAAEc3VtMAIAAAAITGF5ZXIxTjAEAAAABmRlYnVnMAgFAAAACyR0MDE2NzIxNzE4AAAAAl8xBAAAAARzaWcwCAUAAAALJHQwMTY3MjE3MTgAAAACXzIEAAAACyR0MDE3MjMxNzY5CQEAAAAHc2lnbW9pZAAAAAIFAAAABHN1bTECAAAACExheWVyMU4xBAAAAAZkZWJ1ZzEIBQAAAAskdDAxNzIzMTc2OQAAAAJfMQQAAAAEc2lnMQgFAAAACyR0MDE3MjMxNzY5AAAAAl8yCQAFFAAAAAIJAARMAAAAAgUAAAAEc2lnMAkABEwAAAACBQAAAARzaWcxBQAAAANuaWwJAAROAAAAAgUAAAAGZGVidWcwBQAAAAZkZWJ1ZzEBAAAAEWZvcndhcmRQYXNzTGF5ZXIyAAAABAAAAAVpbnB1dAAAAAd3ZWlnaHRzAAAABmJpYXNlcwAAAAtkZWJ1Z1ByZWZpeAQAAAAEc3VtMAkAAGQAAAACCQAAZAAAAAIJAABrAAAAAwkAAZEAAAACBQAAAAVpbnB1dAAAAAAAAAAAAAkAAZEAAAACCQABkQAAAAIFAAAAB3dlaWdodHMAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAPQkAJAABrAAAAAwkAAZEAAAACBQAAAAVpbnB1dAAAAAAAAAAAAQkAAZEAAAACCQABkQAAAAIFAAAAB3dlaWdodHMAAAAAAAAAAAAAAAAAAAAAAAEAAAAAAAAPQkAJAAGRAAAAAgUAAAAGYmlhc2VzAAAAAAAAAAAABAAAAAskdDAyMDM4MjA4NAkBAAAAB3NpZ21vaWQAAAACBQAAAARzdW0wAgAAAAhMYXllcjJOMAQAAAAGZGVidWcwCAUAAAALJHQwMjAzODIwODQAAAACXzEEAAAABHNpZzAIBQAAAAskdDAyMDM4MjA4NAAAAAJfMgkABRQAAAACBQAAAARzaWcwBQAAAAZkZWJ1ZzAAAAABAAAAAWkBAAAAB3ByZWRpY3QAAAACAAAABmlucHV0MQAAAAZpbnB1dDIEAAAADHNjYWxlZElucHV0MQMJAAAAAAAAAgUAAAAGaW5wdXQxAAAAAAAAAAABAAAAAAAAD0JAAAAAAAAAAAAABAAAAAxzY2FsZWRJbnB1dDIDCQAAAAAAAAIFAAAABmlucHV0MgAAAAAAAAAAAQAAAAAAAA9CQAAAAAAAAAAAAAQAAAAGaW5wdXRzCQAETAAAAAIFAAAADHNjYWxlZElucHV0MQkABEwAAAACBQAAAAxzY2FsZWRJbnB1dDIFAAAAA25pbAQAAAALJHQwMjMzNTI0MzMJAQAAABFmb3J3YXJkUGFzc0xheWVyMQAAAAQFAAAABmlucHV0cwUAAAANbGF5ZXIxV2VpZ2h0cwUAAAAMbGF5ZXIxQmlhc2VzAgAAAAZMYXllcjEEAAAADGxheWVyMU91dHB1dAgFAAAACyR0MDIzMzUyNDMzAAAAAl8xBAAAAAtkZWJ1Z0xheWVyMQgFAAAACyR0MDIzMzUyNDMzAAAAAl8yBAAAAAskdDAyNDM4MjU0MgkBAAAAEWZvcndhcmRQYXNzTGF5ZXIyAAAABAUAAAAMbGF5ZXIxT3V0cHV0BQAAAA1sYXllcjJXZWlnaHRzBQAAAAxsYXllcjJCaWFzZXMCAAAABkxheWVyMgQAAAAMbGF5ZXIyT3V0cHV0CAUAAAALJHQwMjQzODI1NDIAAAACXzEEAAAAC2RlYnVnTGF5ZXIyCAUAAAALJHQwMjQzODI1NDIAAAACXzIJAAROAAAAAgkABE4AAAACCQAETAAAAAIJAQAAAAxJbnRlZ2VyRW50cnkAAAACAgAAAAZyZXN1bHQFAAAADGxheWVyMk91dHB1dAUAAAADbmlsBQAAAAtkZWJ1Z0xheWVyMQUAAAALZGVidWdMYXllcjIAAAAAFaLXgQ==", "height": 3082469, "applicationStatus": "succeeded", "spentComplexity": 0 } View: original | compacted Prev: CTK5XZL5t8cmyyJ2vqgR2bTBcRpvXGeA2RrB6hGC8GVB Next: GY2gPoqivX3eeH3zPPj8SdvLMnbisXfuy2r24TQoJGgR Diff:
OldNewDifferences
11 {-# STDLIB_VERSION 5 #-}
22 {-# SCRIPT_TYPE ACCOUNT #-}
33 {-# CONTENT_TYPE DAPP #-}
4-let layer1Weights = [[600497, 600733], [414197, 414253]]
4+let layer1Weights = [[600497, 600732], [414197, 414253]]
55
6-let layer1Biases = [-259050, -635638]
6+let layer1Biases = [-259051, -635637]
77
88 let layer2Weights = [[832965, -897142]]
99
1010 let layer2Biases = [-381179]
1111
1212 func exp_approximation (x) = {
13- let scale = toBigInt(100000)
14- let e = toBigInt(271828)
15- let factor1 = fraction(x, scale, toBigInt(1))
16- let factor2 = fraction(fraction((x * x), scale, toBigInt(1)), toBigInt((2 * 100000)), toBigInt(1))
17- let factor3 = fraction(fraction(((x * x) * x), scale, toBigInt(1)), toBigInt(((6 * 100000) * 100000)), toBigInt(1))
18- fraction((((toBigInt(1000000) + factor1) + factor2) + factor3), toBigInt(1), toBigInt(1))
13+ let scale = 10000
14+ let scaled_x = (x / scale)
15+ let e = 27183
16+ let factor1 = scaled_x
17+ let factor2 = ((scaled_x * scaled_x) / (2 * scale))
18+ let factor3 = (((scaled_x * scaled_x) * scaled_x) / ((6 * scale) * scale))
19+ (((10000 + factor1) + factor2) + factor3)
1920 }
2021
2122
2223 func sigmoid (z,debugPrefix) = {
23- let base = toBigInt(100000)
24- let positiveZ = toBigInt(if ((0 > z))
24+ let base = 10000
25+ let positiveZ = if ((0 > z))
2526 then -(z)
26- else z)
27+ else z
2728 let expValue = exp_approximation(positiveZ)
28- let sigValue = fraction(base, (base + expValue), toBigInt(1))
29- $Tuple2([IntegerEntry((debugPrefix + "positiveZ"), toInt(positiveZ)), IntegerEntry((debugPrefix + "expValue"), toInt(expValue)), IntegerEntry((debugPrefix + "sigValue"), toInt(sigValue))], toInt(sigValue))
29+ let sigValue = ((base * 10000) / (base + expValue))
30+ $Tuple2([IntegerEntry((debugPrefix + "positiveZ"), positiveZ), IntegerEntry((debugPrefix + "expValue"), expValue), IntegerEntry((debugPrefix + "sigValue"), sigValue)], sigValue)
3031 }
3132
3233
3334 func forwardPassLayer1 (input,weights,biases,debugPrefix) = {
3435 let sum0 = ((fraction(input[0], weights[0][0], 1000000) + fraction(input[1], weights[0][1], 1000000)) + biases[0])
3536 let sum1 = ((fraction(input[0], weights[1][0], 1000000) + fraction(input[1], weights[1][1], 1000000)) + biases[1])
36- let $t017241770 = sigmoid(sum0, "Layer1N0")
37- let debug0 = $t017241770._1
38- let sig0 = $t017241770._2
39- let $t017751821 = sigmoid(sum1, "Layer1N1")
40- let debug1 = $t017751821._1
41- let sig1 = $t017751821._2
37+ let $t016721718 = sigmoid(sum0, "Layer1N0")
38+ let debug0 = $t016721718._1
39+ let sig0 = $t016721718._2
40+ let $t017231769 = sigmoid(sum1, "Layer1N1")
41+ let debug1 = $t017231769._1
42+ let sig1 = $t017231769._2
4243 $Tuple2([sig0, sig1], (debug0 ++ debug1))
4344 }
4445
4546
4647 func forwardPassLayer2 (input,weights,biases,debugPrefix) = {
4748 let sum0 = ((fraction(input[0], weights[0][0], 1000000) + fraction(input[1], weights[0][1], 1000000)) + biases[0])
48- let $t020902136 = sigmoid(sum0, "Layer2N0")
49- let debug0 = $t020902136._1
50- let sig0 = $t020902136._2
49+ let $t020382084 = sigmoid(sum0, "Layer2N0")
50+ let debug0 = $t020382084._1
51+ let sig0 = $t020382084._2
5152 $Tuple2(sig0, debug0)
5253 }
5354
6162 then 1000000
6263 else 0
6364 let inputs = [scaledInput1, scaledInput2]
64- let $t023872485 = forwardPassLayer1(inputs, layer1Weights, layer1Biases, "Layer1")
65- let layer1Output = $t023872485._1
66- let debugLayer1 = $t023872485._2
67- let $t024902594 = forwardPassLayer2(layer1Output, layer2Weights, layer2Biases, "Layer2")
68- let layer2Output = $t024902594._1
69- let debugLayer2 = $t024902594._2
65+ let $t023352433 = forwardPassLayer1(inputs, layer1Weights, layer1Biases, "Layer1")
66+ let layer1Output = $t023352433._1
67+ let debugLayer1 = $t023352433._2
68+ let $t024382542 = forwardPassLayer2(layer1Output, layer2Weights, layer2Biases, "Layer2")
69+ let layer2Output = $t024382542._1
70+ let debugLayer2 = $t024382542._2
7071 (([IntegerEntry("result", layer2Output)] ++ debugLayer1) ++ debugLayer2)
7172 }
7273
Full:
OldNewDifferences
11 {-# STDLIB_VERSION 5 #-}
22 {-# SCRIPT_TYPE ACCOUNT #-}
33 {-# CONTENT_TYPE DAPP #-}
4-let layer1Weights = [[600497, 600733], [414197, 414253]]
4+let layer1Weights = [[600497, 600732], [414197, 414253]]
55
6-let layer1Biases = [-259050, -635638]
6+let layer1Biases = [-259051, -635637]
77
88 let layer2Weights = [[832965, -897142]]
99
1010 let layer2Biases = [-381179]
1111
1212 func exp_approximation (x) = {
13- let scale = toBigInt(100000)
14- let e = toBigInt(271828)
15- let factor1 = fraction(x, scale, toBigInt(1))
16- let factor2 = fraction(fraction((x * x), scale, toBigInt(1)), toBigInt((2 * 100000)), toBigInt(1))
17- let factor3 = fraction(fraction(((x * x) * x), scale, toBigInt(1)), toBigInt(((6 * 100000) * 100000)), toBigInt(1))
18- fraction((((toBigInt(1000000) + factor1) + factor2) + factor3), toBigInt(1), toBigInt(1))
13+ let scale = 10000
14+ let scaled_x = (x / scale)
15+ let e = 27183
16+ let factor1 = scaled_x
17+ let factor2 = ((scaled_x * scaled_x) / (2 * scale))
18+ let factor3 = (((scaled_x * scaled_x) * scaled_x) / ((6 * scale) * scale))
19+ (((10000 + factor1) + factor2) + factor3)
1920 }
2021
2122
2223 func sigmoid (z,debugPrefix) = {
23- let base = toBigInt(100000)
24- let positiveZ = toBigInt(if ((0 > z))
24+ let base = 10000
25+ let positiveZ = if ((0 > z))
2526 then -(z)
26- else z)
27+ else z
2728 let expValue = exp_approximation(positiveZ)
28- let sigValue = fraction(base, (base + expValue), toBigInt(1))
29- $Tuple2([IntegerEntry((debugPrefix + "positiveZ"), toInt(positiveZ)), IntegerEntry((debugPrefix + "expValue"), toInt(expValue)), IntegerEntry((debugPrefix + "sigValue"), toInt(sigValue))], toInt(sigValue))
29+ let sigValue = ((base * 10000) / (base + expValue))
30+ $Tuple2([IntegerEntry((debugPrefix + "positiveZ"), positiveZ), IntegerEntry((debugPrefix + "expValue"), expValue), IntegerEntry((debugPrefix + "sigValue"), sigValue)], sigValue)
3031 }
3132
3233
3334 func forwardPassLayer1 (input,weights,biases,debugPrefix) = {
3435 let sum0 = ((fraction(input[0], weights[0][0], 1000000) + fraction(input[1], weights[0][1], 1000000)) + biases[0])
3536 let sum1 = ((fraction(input[0], weights[1][0], 1000000) + fraction(input[1], weights[1][1], 1000000)) + biases[1])
36- let $t017241770 = sigmoid(sum0, "Layer1N0")
37- let debug0 = $t017241770._1
38- let sig0 = $t017241770._2
39- let $t017751821 = sigmoid(sum1, "Layer1N1")
40- let debug1 = $t017751821._1
41- let sig1 = $t017751821._2
37+ let $t016721718 = sigmoid(sum0, "Layer1N0")
38+ let debug0 = $t016721718._1
39+ let sig0 = $t016721718._2
40+ let $t017231769 = sigmoid(sum1, "Layer1N1")
41+ let debug1 = $t017231769._1
42+ let sig1 = $t017231769._2
4243 $Tuple2([sig0, sig1], (debug0 ++ debug1))
4344 }
4445
4546
4647 func forwardPassLayer2 (input,weights,biases,debugPrefix) = {
4748 let sum0 = ((fraction(input[0], weights[0][0], 1000000) + fraction(input[1], weights[0][1], 1000000)) + biases[0])
48- let $t020902136 = sigmoid(sum0, "Layer2N0")
49- let debug0 = $t020902136._1
50- let sig0 = $t020902136._2
49+ let $t020382084 = sigmoid(sum0, "Layer2N0")
50+ let debug0 = $t020382084._1
51+ let sig0 = $t020382084._2
5152 $Tuple2(sig0, debug0)
5253 }
5354
5455
5556 @Callable(i)
5657 func predict (input1,input2) = {
5758 let scaledInput1 = if ((input1 == 1))
5859 then 1000000
5960 else 0
6061 let scaledInput2 = if ((input2 == 1))
6162 then 1000000
6263 else 0
6364 let inputs = [scaledInput1, scaledInput2]
64- let $t023872485 = forwardPassLayer1(inputs, layer1Weights, layer1Biases, "Layer1")
65- let layer1Output = $t023872485._1
66- let debugLayer1 = $t023872485._2
67- let $t024902594 = forwardPassLayer2(layer1Output, layer2Weights, layer2Biases, "Layer2")
68- let layer2Output = $t024902594._1
69- let debugLayer2 = $t024902594._2
65+ let $t023352433 = forwardPassLayer1(inputs, layer1Weights, layer1Biases, "Layer1")
66+ let layer1Output = $t023352433._1
67+ let debugLayer1 = $t023352433._2
68+ let $t024382542 = forwardPassLayer2(layer1Output, layer2Weights, layer2Biases, "Layer2")
69+ let layer2Output = $t024382542._1
70+ let debugLayer2 = $t024382542._2
7071 (([IntegerEntry("result", layer2Output)] ++ debugLayer1) ++ debugLayer2)
7172 }
7273
7374

github/deemru/w8io/6500d08 
49.11 ms