I am using Git to push our application to AWS though Elastic Beanstalk. AWS has created the means to do this, although from my understanding it is not exactly a traditional repository. So, for example, I can't pull
from my Beanstalk project.
I am confused about what is happening during a push
and why sometimes it is pushing much more data than others. My typical process is as follows
//create, edit files
$ git add .
$ git commit -m "message here"
[master 7d38f1f] message here
5 files changed, 27 insertions(+), 19 deletions(-)
$ git aws.push
Counting objects: 5131, done.
Delta compression using up to 8 threads.
Compressing objects: 100% (4930/4930), done.
Writing objects: 100% (5131/5131), 40.70 MiB | 391.00 KiB/s, done.
Total 5131 (delta 1487), reused 0 (delta 0)
remote: processing
To https://exampleLocation
+ 7f9d9dd...7d38f1f HEAD -> master (forced update)
Now maybe I'll edit a few files and do another push. This time it will be much smaller/faster
$ git add .
$ git commit -m "message2"
[master e2f4412] message2
1 file changed, 1 insertion(+)
$ git aws.push
Counting objects: 9, done.
Delta compression using up to 8 threads.
Compressing objects: 100% (5/5), done.
Writing objects: 100% (5/5), 472 bytes | 0 bytes/s, done.
Total 5 (delta 3), reused 0 (delta 0)
remote:
To https//exampleLocation
7d38f1f..e2f4412 HEAD -> master
Tomorrow, when I do my first push, it will might be the big one again...
This question alludes that only changes are sent, and in the above example you can see that to be true. However, tomorrow when I come in to work my first push will likely be similar to the first one. So somebody is losing track of what has been changed. How can I get it to remember?
Thank you