Don’t Trust GPT — Off Topic and Wrong!

I will be demo’ing some AI and Cybersecurity over the next few weeks. So, let’s see if GPT actually has superfical knowledge or not. Let’s…

Photo by Andrea De Santis on Unsplash

Can You Trust GPT — When It’s Off-Topic and Wrong?

I will be demo’ing some AI and Cybersecurity over the next few weeks. So, let’s see if GPT actually has superficial knowledge or not. Let’s ask a fairly trivial question of:

The answer looks to be well written — although rather lacking any type of style. It is bland and just looks like a standard Stack Overflow answer. But, while it seems authoritative, it is poorly described:

In ECC, the public key is a point on an elliptic curve, and the private key is a random number. To encrypt a message, the sender computes a point on the elliptic curve using the receiver’s public key and then uses this point to derive a shared secret key. The sender then uses this shared secret key to encrypt the message, which can only be decrypted by the receiver, who knows the corresponding private key.

For a simple answer, it is acceptable but misses a good deal — as the short answers it gives tend to generalise things. And when it comes to why the security is better, it can only say that the keys are smaller — and which goes off-topic :

And then:

Again, it goes off-topic, and goes into the type of answer you get in Stack Overflow for answers that are not quite right:

I won’t go into detail, but again the symmetric key part of this never appears, and there are too many k values used. In a real life system the sender generates a random value, and the result is not k². If you probe for a little more detail, I will prompt itself to give more detail. Again, looks like it’s straight out the textbooks:

But let’s get it to do some computation:

And the answer is:

This is completely wrong! 14 times 14 is 196. If we take 196 (mod 97) we get 2, but 2³ + 7 (mod 97) is 15. The other points are wrong too!

The correct answers are [here]:

y^2 = x^3 + 7 (mod 97)

[(1, 28), (1, 69), (5, 61), (5, 36), (12, 38), (12, 59), (13, 78),
(13, 19), (14, 61), (14, 36), (17, 19), (17, 78), (20, 21),
(20, 76), (21, 76), (21, 21), (23, 90), (23, 7), (27, 75),
(27, 22), (29, 7), (29, 90), (32, 59), (32, 38), (33, 32),
(33, 65), (35, 69), (35, 28), (36, 54), (36, 43), (44, 5),
(44, 92), (45, 7), (45, 90), (52, 81), (52, 16), (53, 59),
(53, 38), (55, 30), (55, 67), (56, 76), (56, 21), (57, 67),
(57, 30), (60, 45), (60, 52), (61, 69), (61, 28), (62, 43),
(62, 54), (63, 45), (63, 52), (65, 92), (65, 5), (67, 78), (67, 19),
(68, 16), (68, 81), (71, 45), (71, 52), (72, 22), (72, 75), (73, 65),
(73, 32), (74, 81), (74, 16), (78, 61), (78, 36), (82, 30), (82, 67),
(85, 92), (85, 5), (88, 65), (88, 32), (95, 75), (95, 22), (96, 54),
(96, 43)]

For example for (1,28), we get 28² (mod 97) and which is 8, and where 1³+7 is also 8.

So, can it show that it was wrong:

And, so (2,14) is wrong, as GPT has shown.

Conclusions

For the teaching professional, we must all be worried about the power of GPT. Not because of its technical abilities, but in the ability to short-circuit the education process. If you are in Edinburgh on 2 March 2023, why not pop along to: