Data Manipulation

http://d2l.ai/chapter_preliminaries/ndarray.html

1 Like

1.

x = torch.arange(12, dtype=torch.float32).reshape((3,4))
y = torch.tensor([[2.0, 1, 4, 3], [1, 2, 3, 4], [4, 3, 2, 1]])
x,y,x == y,x < y,x > y

(tensor([[ 0., 1., 2., 3.],
[ 4., 5., 6., 7.],
[ 8., 9., 10., 11.]]),
tensor([[2., 1., 4., 3.],
[1., 2., 3., 4.],
[4., 3., 2., 1.]]),
tensor([[False, True, False, True],
[False, False, False, False],
[False, False, False, False]]),
tensor([[ True, False, True, False],
[False, False, False, False],
[False, False, False, False]]),
tensor([[False, False, False, False],
[ True, True, True, True],
[ True, True, True, True]]))


2.

a = torch.arange(1, 6, dtype =torch.float32).reshape((5, 1))
b = torch.arange(1, 3).reshape((1, 2))
a, b

(tensor([[1],
[2],
[3],
[4],
[5]]),
tensor([[1, 2]]))


a + b

tensor([[2., 3.],
[3., 4.],
[4., 5.],
[5., 6.],
[6., 7.]])


a - b

tensor([[ 0., -1.],
[ 1., 0.],
[ 2., 1.],
[ 3., 2.],
[ 4., 3.]])


a * b

tensor([[1.0000, 0.5000],
[2.0000, 1.0000],
[3.0000, 1.5000],
[4.0000, 2.0000],
[5.0000, 2.5000]])


a / b

tensor([[1, 0],
[2, 1],
[3, 1],
[4, 2],
[5, 2]])


a // b

tensor([[1., 0.],
[2., 1.],
[3., 1.],
[4., 2.],
[5., 2.]])


a \ b

File “” , line 1 a \ b ^ SyntaxError : unexpected character after line continuation character


a ** b

tensor([[ 1., 1.],
[ 2., 4.],
[ 3., 9.],
[ 4., 16.],
[ 5., 25.]])

@StevenJokes There is no \ operator in pytorch. It is actually a special chracter in python, also called the “escape” character. Hence the error.

Let me know if this is not clear.

I have got it from the doc, but thanks anyway.
I’m a new bee of pytorch.

a % b
tensor([[0., 1.],
[0., 0.],
[0., 1.],
[0., 0.],
[0., 1.]])

When having PyTorch selected:

2.1.5. Saving Memory §3:

Fortunately, performing in-place operations in MXNet is easy.

Is it intentional to discuss MXNet even though PyTorch is selected for the code examples?

It is understandable. :sweat_smile:
The original examples are built by MXNet. :joy:( created by mli)

Allow me to point out a small error in Section 2.1.2
“For stylistic convenience, we can write x.sum() as np.sum(x) .” should not appear in PyTorch version because it is not possible to run np.sum(x) if x is a PyTorch tensor.

x = torch.arange(12)
np.sum(x)
---------------------------------------------------------------------------
TypeError                                 Traceback (most recent call last)
<ipython-input-30-1393831a87e1> in <module>
      1 x = torch.arange(12)
----> 2 np.sum(x)

Thanks @hehao98 for pointng that out. We have already fixed that line in this commit and it will be updated with our next release.

98/5000
Is it only possible to use broadcast when the two arrays have a dimension has value equal to one?

Hi @jairo.venegas, no need to be one. You can broadcast anything :wink: This example may give you more idea!

in 2.1.4. Indexing and Slicing

X[0:2, :]
this code suppose to take the 1th and 2nd rows, but why it isn’t typed like that -> X[0:1,:]

Hi @omarkhaled850, i am not fully understand your question. Could you elaborate more on that?

thanks for responding
in the slicing code. we want to take the 1th row (index = 0) and the 2nd row (index = 1) in the example. put the code start from index 0 to index 2 (0:2).
shouldn’t it be (0:1)

Slicing is an indexing syntax that extracts a portion from the tensor. X[m:n] returns the portion of X :

  • Starting with position m
  • Up to but not including n

thanks man, "Up to but not including n" is the key that i was looking for

Hello guys,
In section 2.1.6 [Conversion to Other Python Objects], what do you mean by the line in bold; Unfortunately I can’t get it and I need help. Thanks

Converting to a NumPy tensor, or vice versa, is easy. The converted result does not share memory. This minor inconvenience is actually quite important: when you perform operations on the CPU or on GPUs, you do not want to halt computation, waiting to see whether the NumPy package of Python might want to be doing something else with the same chunk of memory.

Does it mean NumPy and PyTorch are completely independent and may have conflicts with each other?

Edit: According to PyTorch documentation:

Converting a torch Tensor to a NumPy array and vice versa is a breeze. The torch Tensor and NumPy array will share their underlying memory locations, and changing one will change the other.

It says this conversion shares the memory but you said not! Am I right?

This is probably carried over from the MXNET version, where the corresponding operation does create a copy:

asnumpy () : Returns a numpy.ndarray object with value copied from this array.

@anirudh should be able to confirm/deny this.

Thanks, @Aaron and @gphilip for raising this. Most part of the book has common text and we are trying to fix issues like these where the frameworks differ in design. Feel free to raise any other issues if you find something similar in other sections on the forum or the Github repo. Really appreciate it!

This will be fixed in the next release.

1 Like

e = torch.arange(12).reshape(2, -1, 6)
f = torch.tensor([1, 2, 3, 4]).reshape(-1, 4, 1)
e, f, e.shape, f.shape
(tensor([[[ 0, 1, 2, 3, 4, 5]],

    [[ 6,  7,  8,  9, 10, 11]]]), tensor([[[1],
     [2],
     [3],
     [4]]]), torch.Size([2, 1, 6]), torch.Size([1, 4, 1]))

e + f
tensor([[[ 1, 2, 3, 4, 5, 6],
[ 2, 3, 4, 5, 6, 7],
[ 3, 4, 5, 6, 7, 8],
[ 4, 5, 6, 7, 8, 9]],

    [[ 7,  8,  9, 10, 11, 12],
     [ 8,  9, 10, 11, 12, 13],
     [ 9, 10, 11, 12, 13, 14],
     [10, 11, 12, 13, 14, 15]]])